Background: Traffic-related air pollution (TRAP) is a major contributor to urban pollution and varies sharply at the street level, posing a challenge for air quality modeling. Traditional land use regression models combined with data from fixed monitoring stations may be unable to predict and characterize fine-scale TRAP, especially in complex urban environments influenced by various features. This study aims to estimate fine-scale (50 m) concentrations of nitrogen oxides (NO and NO₂) in Hong Kong using a deep learning (DL) structured model.
Methods: We collected data from mobile air quality sensors on buses and crowd-sourced Google real-time traffic status as a proxy for real-time traffic emissions. Our DL model was compared with existing machine learning models to assess performance improvements. Using an interpretable machine learning method, we hierarchically evaluated the global, local, and interaction effects for different features.
Results: Our DL model outperformed existing machine learning models, achieving R values of 0.72 for NO and 0.69 for NO₂. The incorporation of traffic status as a key predictor improved model performance by 9% to 17%. The interpretable machine learning method revealed the importance of traffic-related features and their pairwise interactions.
Conclusion: The results indicate that traffic-related features significantly contribute to TRAP and provide insights and guidance for urban planning. By incorporating crowd-sourced Google traffic information, we assessed traffic abatement scenarios that could inform targeted strategies for improving urban air quality.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.envint.2024.108992 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!