Priyanka
deSouza
*ab,
Karoline
Barkjohn
c,
Andrea
Clements
c,
Jenny
Lee
d,
Ralph
Kahn
e,
Ben
Crawford
f and
Patrick
Kinney
g
aDepartment of Urban and Regional Planning, University of Colorado Denver, Denver, CO 80202, USA. E-mail: priyanka.desouza@ucdenver.edu
bCU Population Center, University of Colorado Boulder, Boulder, CO 80302, USA
cOffice of Research and Development, US Environmental Protection Agency, 109 T.W. Alexander Drive, Research Triangle Park, NC 27711, USA
dDepartment of Biostatistics, Harvard T.H. Chan School of Public Health, Boston, MA 02115, USA
eNASA Goddard Space Flight Center, Greenbelt, MD 20771, USA
fDepartment of Geography and Environmental Sciences, University of Colorado Denver, 80202, USA
gBoston University School of Public Health, Boston, MA 02118, USA
First published on 3rd February 2023
Low-cost sensors (LCSs) are increasingly being used to measure fine particulate matter (PM2.5) concentrations in cities around the world. One of the most commonly deployed LCSs is PurpleAir with ∼15000 sensors deployed in the United States, alone. PurpleAir measurements are widely used by the public to evaluate PM2.5 levels in their neighborhoods. PurpleAir measurements are also increasingly being integrated into models by researchers to develop large-scale estimates of PM2.5. However, the change in sensor performance over time has not been well studied. It is important to understand the lifespan of these sensors to determine when they should be serviced or replaced, and when measurements from these devices should or should not be used for various applications. This paper fills this gap by leveraging the fact that: (1) each PurpleAir sensor is composed of two identical sensors and the divergence between their measurements can be observed, and (2) there are numerous PurpleAir sensors within 50 meters of regulatory monitors allowing for the comparison of measurements between these instruments. We propose empirically-derived degradation outcomes for the PurpleAir sensors and evaluate how these outcomes change over time. On average, we find that the number of ‘flagged’ measurements, where the two sensors within each PurpleAir sensor disagree, increases with time to ∼4% after 4 years of operation. Approximately, 2 percent of all PurpleAir sensors were permanently degraded. The largest fraction of permanently degraded PurpleAir sensors appeared to be in the hot and humid climate zone, suggesting that sensors in these locations may need to be replaced more frequently. We also find that the bias of PurpleAir sensors, or the difference between corrected PM2.5 levels and the corresponding reference measurements, changed over time by −0.12 μg m−3 (95% CI: −0.13 μg m−3, −0.10 μg m−3) per year. The average bias increases dramatically after 3.5 years. Further, climate zone is a significant modifier of the association between degradation outcomes and time.
Environmental significanceLow-cost air quality sensors are widely used to fill in air quality monitoring gaps. However, little is known about the performance of these sensors over time. We evaluate degradation patterns across a network of widely-used low-cost particulate matter sensors, the PurpleAir. Overall, we find that 4% of measurements from PurpleAir sensors are degraded after 4 years of operation. Rates of degradation vary by climate zone. We also identify permanently degraded PurpleAir sensors that should be removed from operation. Our work provides a framework to quantify degradation in other low-cost air quality sensors. |
Low-cost sensors (LCSs) (<$2500 USD as defined by the U.S. EPA15) have the potential to capture concentrations of particulate matter (PM) in previously unmonitored locations and democratize air pollution information.13,16–21 Measurements from these devices are increasingly being integrated into models to develop large-scale exposure assessments.22–24
Most low-cost PM sensors rely on optical measurement techniques that introduce potential differences in mass estimations compared to reference monitors (i.e., FRM/FEM monitors).25–27 Optical sensor methods do not directly measure mass concentrations; rather, they measure light scattering of particles having diameters typically > ∼0.3 μm. Several assumptions are typically made to convert light scattering into mass concentrations that can introduce errors in the results. In addition, unlike reference monitors, LCSs do not dry particles before measuring them, so PM concentrations reported by LCSs can be biased high due to particle hygroscopic growth of particles when ambient relative humidity (RH) is high. Many research groups have developed different techniques to correct the raw LCS measurements from PM sensors. These models often include environmental variables, such as RH, temperature (T), and dewpoint (D), as predictors of the ‘true’ PM concentration.
However, little work has been done to evaluate the performance of low-cost PM sensors over time. There is evidence that the performance of these instruments can be affected by high PM events which can also impact subsequent measurements if the sensors are not cleaned properly.28 Although there has been some research evaluating drift in measurements from low-cost electrochemical gas sensors,29,30 there has been less work evaluating drift and degradation in low-cost PM sensors and identifying which factors affect these outcomes. An understanding of degradation could lead to better protocols for correcting low-cost PM sensors and could provide users with information on when to service or replace their sensors or whether data should or should not be used for certain applications.
This paper evaluates the performance of the PurpleAir sensor, one of the most common low-cost PM sensors over time. We chose to conduct this analysis with PurpleAir because:
(1) There is a sizable number of PurpleAir sensors within 50 meters of regulatory monitors that allows for comparison between PurpleAir measurements and reference data over time, and
(2) Each PurpleAir sensor consists of two identical PM sensors making it possible to evaluate how the two sensors disagree over time, and the different factors that contribute to this disagreement.
(3) Several studies have evaluated the short-term performance of the PurpleAir sensors at many different locations, under a variety of conditions around the world.31,32 However, none of these studies has evaluated the performance of the PurpleAir sensors over time. We aim to fill this gap.
The PA-II PurpleAir sensor operates for 10 s at alternating intervals and provides 2 min averaged data (prior to 30 May 2019, this was 80 s averaged data). The Plantower sensor components measure light scattering with a laser at 680 ± 10 nm wavelength 33,34 and are factory calibrated using ambient aerosol across several cities in China.27 The Plantower sensor reports estimated mass concentrations of particles with aerodynamic diameters < 1 μm (PM1), < 2.5 μm (PM2.5), and < 10 μm (PM10). For each PM size fraction, the values are reported in two ways, labeled cf_1 and cf_atm, in the PurpleAir dataset, which match the “raw” Plantower outputs.
The ratio of cf_atm and cf_1 (i.e. [cf_atm]/[cf_1]) is equal to 1 for PM2.5 concentrations below 25 μg m−3 (as reported by the sensor) and then transitions to a two-thirds ratio at a higher PM concentration (cf_1 concentrations are higher). The cf_atm data, displayed on the PurpleAir map, are the lower measurement of PM2.5 and are referred to as the “raw” data in this paper when making comparisons between initial and corrected datasets.33 When a PurpleAir sensor is connected to the internet, data are sent to PurpleAir's data repository. Users can choose to make their data publicly viewable (public) or control data sharing (private). All PurpleAir sensors also report RH and T levels.
For this study, data from 14927 PurpleAir sensors operating in the United States (excluding US territories) between 1 January 2017 to 20 July 2021 were downloaded from the API at 15-minute time resolution. A small number of PurpleAir sensors were operational before 2017. However, given that the number of PurpleAir sensors increased dramatically from 2017 onwards, we choose January 1 2017 as the start date of our analysis. Overall, 26.2% of dates had missing measurements, likely due to power outages or loss of WiFi that prevented the PurpleAir sensors from transmitting data. Of the sensors in our dataset, 2989 were missing channel B data, leaving us with 483511216 measurements from 11938 sensors with both channel A and B data. We removed all records with missing PM2.5 measurements in cf_1 channels A and B (∼0.9% of the data). We then removed all records with missing T and RH data (∼2.6% of all data). Of the non-missing records, all measurements where PM2.5 in cf._1 channels A and B were both >1500 μg m−3 were removed, as they correspond to conditions beyond the operating range of the PurpleAir sensor.25 We also removed measurements where T was ≤ −50 °C or ≥ 100 °C, or where RH was >99%, as these corresponded to extreme conditions (∼4.2% of all records). The remaining dataset contained 457488977 measurements from 11933 sensors.
The 15-minute data were averaged to 1 h intervals. A 75% data completeness threshold was used (at least 3 15minute measurements in an hour) based on channels A and B. This methodology ensured that the averages used were representative of hourly averages. We defined the hourly mean PM2.5 cf_1 as the average of the PM2.5 cf_1 measurements from channels A and B. We defined hourly mean PM2.5 cf_atm as the average of PM2.5 cf_atm measurements from channels A and B. We also calculated hourly mean T and RH from the 15 min averaged data from each PurpleAir sensor.
Overall, the dataset included 114259940 valid hourly averaged measurements with non-missing PM2.5 data in channels A or B corresponding to 11932 PurpleAir sensors (8312155 measurements from 935 indoor sensors and 105947785 measurements from 10997 outdoor sensors). A description of the number of sensors and measurements by the state is provided in Table S1 in the ESI.† (Fig. S1 in the ESI† displays the locations of indoor and outdoor PurpleAir sensors.) Of the 11932 PurpleAir sensors, 1377 (∼11.5%) had stopped reporting data at least a day before the data were downloaded (i.e., 20 July 2021), whereas the remaining sensors were still in operation (Fig. 1).
If there was more than one reference monitor within 50 meters of a PurpleAir sensor, measurements were retained from one of the reference monitors. We prioritized retaining data from reference monitors that did not rely on light scattering techniques as these instruments tend to have an additional error when estimating aerosol mass.35
From the resulting dataset, we found that the Pearson correlation coefficient (R) between mean PM2.5 cf_1 and reference PM2.5 concentrations was 0.86, whereas the correlation between PM2.5 cf_atm and reference PM2.5 concentrations was 0.83. Henceforth, when describing PurpleAir measurements, we consider only the mean PM2.5 cf_1 concentrations.
A data-driven approach was adopted to determine if we should use a similar threshold in this study. We flagged measurements where the Δ > 5 μg m−3 and when the % difference between channels A and B was greater than the top percentile of the distribution of the % difference between A and B channels for each PurpleAir sensor. We allowed the percentile threshold to range from 0.0 to 0.99, by increments of 0.01. We use percentiles as a threshold instead of standard deviation as the % difference between the A and B channels is not normally distributed. At each step, we then compared the unflagged PurpleAir measurements with the corresponding reference data using the metrics: Pearson correlation coefficient (R) and the normalized root mean squared error (nRMSE). The percentile threshold that led to the best agreement between the PurpleAir sensor and the corresponding reference monitor was chosen. We calculated nRMSE in this study by normalizing the root mean square error (RMSE) by the standard deviation of PM2.5 from the corresponding reference monitor. As a sensitivity test, we repeated the above analysis after removing records where the reference monitor relied on a light scattering technique (namely the Teledyne and the Grimm instruments), thus eliminating the more error-prone data (Fig. S3†). We note that past studies have shown that the Beta-Attenuation Mass Monitors (BAM) are likely to experience more noise at low PM2.5 concentrations.35,36
After determining the threshold to flag measurements using the collocated data (Fig. 2), we evaluated the number of flagged measurements for each of the 11932 PurpleAir sensors in our sample. We propose the percentage of flagged measurements at a given operational time (from the time, in hours, since each sensor started operating) as a potential degradation outcome. To visually examine if a threshold value existed beyond which these outcomes increased significantly, we plotted this outcome as well as the percentage of cumulative flagged measurements over time (Fig. 3). We evaluated whether the distribution of PM2.5, RH, and T conditions for flagged measurements is statistically different from that for unflagged measurements (Table 2).
Unflagged data (n = 112716535, 99%) | Flagged data (n = 1543405, 1%) | |
---|---|---|
Raw mean PM2.5 (mean of channel a and channel B) (μg m−3) | Min/max: 0/1459 | Min/max: 2.5/1339 |
Mean: 10 | Mean: 26 | |
Median: 5 | Median: 14 | |
1st Quartile: 2 | 1st Quartile: 7 | |
3rd Quartile: 11 | 3rd Quartile: 27 | |
RH (%) | Min/max: 0/99 | Min/max: 0/99 |
Mean: 46 | Mean: 43 | |
Median: 48 | Median: 44 | |
1st Quartile: 34 | 1st Quartile: 30 | |
3rd Quartile: 59 | 3rd Quartile: 57 | |
Temperature (°C) | Min/max: −42/68 | Min/max: −46/89 |
Mean: 18 | Mean: 19 | |
Median: 18 | Median: 19 | |
1st Quartile: 11 | 1st Quartile: 13 | |
3rd Quartile: 24 | 3rd Quartile: 26 | |
Month | Jan: 10233928 (98.5%) | Jan: 157728 (1.5%) |
Feb: 9650954 (98.4%) | Feb: 156615 (1.6%) | |
March: 10979861 (98.7%) | March: 141003 (1.3%) | |
April: 10989824 (98.9%) | April: 125060 (1.1%) | |
May: 11671186 (98.8%) | May: 143421 (1.2%) | |
June: 11674808 (98.6%) | June: 160317 (1.4%) | |
July: 9555217 (98.6%) | July: 140255 (1.4%) | |
Aug: 5246854 (98.7%) | Aug: 67196 (1.3%) | |
Sep: 6248360 (98.6%) | Sep: 86200 (1.4%) | |
Oct: 8025096 (98.8%) | Oct: 99753 ((1.2%) | |
Nov: 8759251 (98.6%) | Nov: 120721 (1.4%) | |
Dec: 9681196 (98.5%) | Dec: 145136 (1.5%) |
For each PurpleAir sensor, at each operational hour, we evaluated the percentage of flagged hourly averages at the given hour and for all subsequent hours. We designated a PurpleAir sensor as permanently degraded if more than 40% of the current and subsequent hourly averages were flagged and the sensor operated for at least 100 hours after the current hour (Fig. 4; Fig. S4†). In sensitivity analyses, we evaluated the number of PurpleAir sensors that would be considered ‘degraded’ for different thresholds (Fig. S5†). We also examined where such sensors were deployed.
A limitation of using the percentage of flagged measurements as a degradation metric is that it does not account for the possibility that channels A and B might both degrade in a similar manner. Therefore, we rely on a second approach, using collocated reference monitoring measurements, to evaluate this aspect of possible degradation.
PM2.5, reference = PM2.5 × s1 + RH × s2 + b + ε | (1) |
Here, PM2.5,reference is the reference monitor measurement; PM2.5 is the PurpleAir measurement calculated by averaging concentrations reported by channels A and B; RH is the relative humidity reported by the PurpleAir sensor. We empirically derived coefficients: s1, s2, and b by regressing uncorrected PurpleAir PM2.5 measurements on reference measurements of PM2.5. ε denotes error from a standard normal distribution. We evaluated one correction model for all PurpleAir sensors in our dataset in a similar manner to Barkjohn et al. (2021). We evaluated and plotted the correction error, which is defined as the difference between the corrected measurement and the corresponding reference PM2.5 measurement in μg m−3. In supplementary analyses, we repeat this process using nine additional correction functions ranging from simple linear regressions to more complex machine learning algorithms, some of which additionally correct for T and D, in addition to RH (Table S3†), to evaluate the sensitivity of our results to the correction model used. A key concern is that some part of the correction error observed might not be due to degradation but to inadequate correction of RH or other environmental parameters. We plot correction error versus RH to visually assess if such a dependence exists. Some of the supplementary correction models used to rely on non-linear corrections for RH. Research has shown that a non-linear correction equation might be more suitable to correct for PurpleAir measurements above ∼500 μg m−3 of PM2.5 levels.37 The machine learning models that we used in the supplement can identify such patterns using statistical learning. A full description of these additional models can be found in deSouza et al. (2022).25
Degradation outcome = f + d × hour of operation + ε | (2) |
where f denotes a constant intercept; d denotes the association between operational time (number of hours since each sensor was deployed) and the degradation outcome as the percentage of (cumulative) flagged measurements over all PurpleAir sensors at a given operational time; and ε denotes error from a standard normal distribution.
For the degradation outcomes under consideration, we evaluated whether the associations were different in subgroups stratified by IECC Climate Zones that represent different T and RH conditions (Table S2† contains information on the number of PurpleAir measurements and reported PM2.5 concentrations by climate zone.) When evaluating the impact of climate zone on the percentage of flagged measurements, we examined the impact on outside devices alone, as indoor environments may not always reflect outside conditions due to heating, cooling, general sheltering, etc. Note that when joining climate zones with the complete dataset of PurpleAir IDs, there were a handful of sensors which did not fall within a climate zone. (This was not the case for our subset of collocated PurpleAir sensors.) We removed data corresponding to these sensors when evaluating climate zone-specific associations, corresponding to 2.9% of all data records Fig. S2 in the ESI† shows where these sensors were located.
We also tested whether the cumulative number of PM2.5 measurements recorded over 50, 100, and 500 μg m−3 by individual PurpleAir sensors significantly modifies the association between operational time and the correction error, as previous work has found that low-cost optical PM sensors can degrade after exposure to high PM concentrations.28 As the correction error will be larger at higher PM2.5 concentrations,25,38 we also evaluated this association after normalizing the correction error by (PM2.5,corrected + PM2.5,reference)/2 to make it easier to interpret how cumulative exposure to high PM2.5 measurements can affect the association between degradation and hour of operation.
The merged PurpleAir and reference measurements dataset only included measurements from outdoor PurpleAir sensors. We also evaluated the indoor/outdoor-specific associations between the percentage of flagged measurements and hours of operation.
Finally, we tested for potential non-linearities between the degradation outcomes under consideration and the time of operation. Penalized splines (p-splines) were used to flexibly model the associations between the error and time of operation using a generalized additive model [GAM; degradation outcome ∼s(hour)]. We used a generalized cross-validation (GCV) criterion to select the optimal number of degrees of freedom (df) and plotted the relationships observed. Ninety-five percent confidence intervals (CIs) were evaluated by m-out-n bootstrap, which creates non-parametric CIs by randomly resampling the data. Briefly, we selected a bootstrapped sample of monitors, performed the correction, and then fit GAMs in each bootstrap sample using sensor ID clusters (100 replicates; Fig. 6).
All analyses were conducted using the software R. In all analyses, p-values <0.05 were taken to represent statistical significance.
When we repeated this analysis excluding measurements from reference monitors that relied on light scattering techniques, using the 86th percentile yielded marginally better results (the metrics differed by < 1%) than using the 85th percentile (Fig. S3 in the ESI†). Given the small difference in results, the 85th percentile is used as the threshold in this study to define a flagged PurpleAir measurement.
Using t-tests, we find that the mean of PM2.5, T, and RH measurements were statistically different (p < 0.05) for flagged PurpleAir measurements compared to unflagged measurements (Table 2). PM2.5 and T measurements recorded when a measurement was flagged were higher than for unflagged measurements, whereas RH tended to be lower. The differences between RH and T values for flagged versus non-flagged measurements are small. The difference in PM2.5 distribution was due in part to the way flags have been defined. As data are flagged only if concentrations differ by at least 5 μg m−3 different, the minimum average flagged concentration is 2.5 μg m−3 (e.g., A = 0, B = 5). There are no notable differences between the percentage of flagged measurements made every month.
We next evaluated the number of PurpleAir measurements that were permanently degraded, or that had a cumulative mean of flags over subsequent hours of operation ≥ 0.4 for at least 100 hours of operation (i.e., at least 40% of measurements flagged) (Fig. 4). Table 3 displays the fraction of permanently degraded sensors in different climate zones and different locations (inside/outside). It appears that the largest fraction of degraded sensors occurred in the south-east United States, a hot and humid climate. Fig. S4† displays the cumulative mean of flag for each ‘permanently degraded’ sensor (the title of each plot corresponds to the sensor ID as provided on the PurpleAir website) at each instance of time. Fig. S4† also depicts the starting year of each permanently degraded sensor. The sensor age varied widely over the set of permanently degraded sensors, indicating that permanent degradation is not dictated by time dependence.
Percentage of permanently degraded sensors | |
---|---|
All | 240 out of 11932 (2.0%) |
Device location | |
Inside | 2 out of 935 (0.21%) |
Outside | 238 out of 10997 (2.2%) |
Climate zone | |
Cold | 51 out of 2458 (2.1%) |
Hot-dry | 54 out of 2680 (2.0%) |
Hot-humid | 11 out of 281 (3.9%) |
Marine | 84 out of 4842 (1.7%) |
Mixed-dry | 3 out of 361 (0.8%) |
Mixed-humid | 24 out of 750 (3.2%) |
Subarctic | 1 out of 58 (1.7%) |
Very cold | 3 of 108 (2.8%) |
No information | 9 of 394 (2.3%) |
Note that from Fig. S4† some of the 240 sensors identified appear to recover or behave normally after a long interval (>100 hours) of degradation (cumulative mean of flag decreases). This could be an artifact of the way the cumulative mean of the flagged indicator is calculated. If the final few measurements of the sensor are not flagged, then the cumulative mean for the final hours of operation of the PurpleAir sensors might be low. It is also possible that some of the sensors could have been temporarily impacted by dust or insects. The owner of the PurpleAir sensors might have cleaned the instruments or replaced the internal Plantower sensors or cleaned out the sensors which could have caused the sensors to recover.
Fig. S5A and S5B† are maps showing locations of PurpleAir sensors that had a cumulative mean of ‘flag’ over subsequent hours of operation of ≥ 0.3 (number of sensors = 323) and 0.5 (number of sensors = 182), respectively, for at least 100 hours of operation.
The correction error (PM2.5,corrected–PM2.5,reference) appeared to become negatively biased over time: −0.12 (95% CI −0.13, −0.10) μg m−3 per year of operation, except for sensors in hot and dry environments where the error was positively biased and increased over time by 0.08 (95% CI: 0.06, 0.09) μg m−3 per year of operation. Wildfires often occur in hot-dry environments. Research has shown that the correction approach could overcorrect the PurpleAir measurements at very high smoke concentrations, potentially explaining the disagreement between the corrected PurpleAir and reference measurements in these environments.39 We note that mean PM2.5 concentrations were highest in hot-dry environments (Table S2†). In addition, the number of PM2.5 concentrations >100 μg m−3 recorded was the highest in hot-dry environments. The magnitude of the correction error bias over time appears to be highest in hot and humid environments corresponding to −0.92 (95% CI −1.10, −0.75) μg m−3 per year. RH has an impact on PurpleAir performance and can also cause the electronic components inside the sensors to degrade quickly, so it is not altogether surprising that degradation appears to be highest in hot and humid environments. We observed similar results when regressing the correction errors derived using other correction model forms (Table S4†). Climate zone is a significant modifier of the association between both degradation outcomes and time (Table 4).
Associations (95% confidence interval) | ||
---|---|---|
Dataset | Percentage of flagged measurements | Correction error |
a (*p < 0.05). | ||
All | 0.93* (0.91, 0.94) | −0.12* (−0.13, −0.10) |
Device location | ||
Inside | −0.10* (−0.12, −0.09) | — |
Outside | 1.06* (1.05, 1.08) | — |
Climate zone (outside devices only) | ||
Cold | 0.74* (0.71, 0.76) | −0.27* (−0.29, −0.25) |
Hot-dry | 2.09* (2.07, 2.12) | 0.08* (0.06, 0.09) |
Hot-humid | 0.34* (0.32, 0.37) | −0.92* (−1.10, −0.75) |
Marine | 0.41* (0.39, 0.44) | −0.13* (−0.15, −0.10) |
Mixed-dry | −0.05* (−0.07, −0.02) | −0.31* (−0.40, −0.21) |
Mixed-humid | 0.54* (0.51, 0.57) | −0.28* (−0.33, −0.23) |
Sub arctic | −0.18* (−0.22, −0.14) | — |
Very cold | 0.13* (0.10, 0.16) | — |
The cumulative number of PM2.5 measurements recorded over 50, 100, and 500 μg m−3 modifies the association between operational time and the correction error significantly, in the negative direction (Table S5†), meaning that sensors that experience more high concentration episodes are more likely to underestimate PM2.5. The increase in the negative bias of the corrected sensor data could be because the absolute magnitude of the correction error will be higher in high PM2.5 environments. When we evaluated the impact of the cumulative number of high PM2.5 measurements on the association between the normalized correction error and operation hour (hours since deployment), we found that the cumulative number of high PM2.5 measurements was not a significant effect modifier of this association (Table S6†). In other words, we did not observe sensors in higher PM2.5 environments degrading faster.
Flagged measurements were more likely to be observed at higher PM2.5 concentrations, lower RH levels, and higher T levels (Table 1). When we evaluated associations between the percentage of flagged measurements and year of operation for sensors in different locations (i.e., outdoor vs. indoor), we found that outdoor sensors degrade much faster than indoor sensors (Table 3). As T and RH impact the likelihood of observing a flagged measurement, this could be because environmental conditions of indoor environments (T and RH) are more regulated than outdoor environments, and indoor instruments tend to be more protected. Our results indicate that the percentage of flagged measurements for indoor environments decreases over time. This could be because of the high percentage of flagged measurements observed in the first 20 hours of operation, and the lack of large changes in the percentage of flagged measurements in later hours of operation in comparison to outdoor sensors. We also note that there is a much smaller number of indoor sensors compared to outdoor instruments (935 compared to 10997), and thus far fewer measurements available, especially at long operational time intervals.
For outdoor sensors, we found that the climate zone in which the sensor was deployed is an important modifier of the association between the percentage of flagged measurements and time. Outdoor sensors in hot-dry climates degrade the fastest, with the percentage of flagged measurements increasing by 2.09% (95% CI 2.07%, 2.12%) every year, an order of magnitude faster than any other climate zone (Table 3). This suggests that on average, outdoor sensors in hot-dry climates likely need to be serviced after ∼3 years, faster than PurpleAir sensors deployed elsewhere.
There was a small number of PurpleAir sensors (240 out of 11932) that were permanently degraded (the cumulative mean of subsequent measurements had over 40% degraded measurements for at least 100 hours). The list of permanently degraded PurpleAir IDs is presented in Fig. S4.† These sensors should be excluded when conducting analyses. The largest fraction of permanently degraded PurpleAir sensors appeared to be in the hot and humid climate zone indicating that sensors in these climates likely needed to be replaced sooner than in others (Table 2). There was no significant relationship between sensor age and permanent degradation, indicating that there may be other factors responsible for causing permanent failure among the PurpleAir sensors. For example, anecdotal evidence suggests that the PurpleAir sensors can be impacted by dust or even insects and degrade the internal components of one or the other PurpleAir channels.
When evaluating the time dependence of the correction error, we found that the PurpleAir instrument bias changes by −0.12 (95% CI: −0.13, −0.10) μg m−3 per year of operation. However, the low associations indicate that this bias is not of much consequence to the operation of PurpleAir sensors. Climate zone was a significant effect modifier of the association between bias and time. The highest associations were observed in hot and humid regions corresponding to −0.92 (95% CI −1.10, −0.75) μg m−3 per year. Exposure to a cumulative number of high PM2.5 measurements did not significantly affect the association between the normalized correction error over time.
It is not altogether surprising that the correction error increases most rapidly in hot and humid climate zones, as past evidence suggests that the performance of PurpleAir is greatly impacted by RH. It is surprising that this is not the case for the other degradation outcomes considered in this study: % of flagged measurements. It is likely that the percentage of flagged measurements increases most rapidly over time in hot and dry environments because such environments tend to be dusty and dust can degrade fan performance and accumulate in the air flow path and optical components of the PurpleAir sensors which can lead to disagreement between the two Plantower sensors. We note that under conditions of wildfire smoke, also prevalent in hot and dry climates, the calibration error could also be magnified due to under-correction of the PurpleAir data. Future work is needed to evaluate the impact of wildfire-smoke on the performance of PurpleAir sensors.
When accounting for non-linearities in the relationship between the correction error and time, Fig. 6a indicates that the bias in the correction error is not linear with time; rather it increases significantly after 30000 hours or 3.5 years. Overall, we found that more work is needed to evaluate degradation in PurpleAir sensors after 3.5 years of operation, due to a paucity of longer-running sensors in the database. Importantly, the degradation outcomes derived in this paper can be used to remove ‘degraded’ PurpleAir measurements in other analyses. We also show that concerns about degradation are more important in some climate zones than others, which may necessitate appropriate cleaning or other maintenance procedures for sensors in different locations.
Footnote |
† Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d2ea00142j |
This journal is © The Royal Society of Chemistry 2023 |