This thesis evaluates different methods for applying the Kalman filter to correct air quality model forecasts in Hong Kong. It introduces two alternative techniques - the logarithmic ratio bias method and relative difference bias method - to model bias beyond the traditional difference method. The performance of each correction method is assessed for PM2.5, NO2 and O3 at three monitoring stations over three years of hourly data using various statistical metrics. The aim is to determine the optimal bias modeling approach for each pollutant and location.
PetroSync - Advanced Seismic Data Acquisition and ProcessingPetroSync
This document provides information about an upcoming training course on advanced seismic data acquisition and processing. The 5-day course will cover key principles in data acquisition, survey design and execution, data preparation and pre-processing, advanced data processing techniques, and data interpretation. It will provide a holistic look at seismic acquisition and processing workflows and include case studies. The course aims to help participants implement optimal workflows, ensure quality control, and make business decisions regarding seismic surveys. It is intended for exploration geophysicists, interpreters, managers, and others involved in seismic data acquisition and processing.
ALTERNATE METHOD TO COMBINE MONITORED AND PREDICTED CONCENTRATIONS IN DISPERS...Sergio A. Guerra
The advent of the short term NAAQS has prompted us to reassess the level of conservatism commonly used in dispersion modeling analyses. An area of conservatism in NAAQS demonstrations relates to the combining of predicted (modeled) concentrations from AERMOD with observed (monitored) concentrations. Normally, some of the highest monitored observations are combined with AERMOD results yielding a very conservative total concentration. For example, in the case of the 1-hour NO2 NAAQS, the chances of the 98th percentile monitored concentration occurring at the same time as the meteorology to generate the 98th percentile ambient concentration is extremely low. Therefore, assuming that both of these happen at the same time is overly conservative.
This presentation includes justification for the use of a reasonable background concentration to combine with the AERMOD predicted concentration. The use of this method, if accepted by regulatory agencies, can help facilities demonstrate compliance in dispersion modeling analyses by assuming a more reasonable background concentration while, at the same time being protective of the NAAQS.
The Powerpoint presentation discusses about the Introduction to CFD and its Applications in various fields as an Introductory topic for Mechanical Engg. Students in General.
This document describes algorithms for detecting single radio pulses in real-time using graphics processing units (GPUs). It presents two new algorithms that use incomplete sets of boxcar filters to detect pulses at accelerated speeds with minimal signal loss. The algorithms were tested on simulated data and were found to process data 266-500 times faster than real-time on GPUs, detecting pulses with a mean 7% reduction in signal power.
Pairing aermod concentrations with the 50th percentile monitored valueSergio A. Guerra
Presentation delivered to the Background Concentrations Workgroup for Air Dispersion Modeling organized by the Minnesota Pollution Control Agency. delivered on March 25, 2014. Three topics covered include 1) Screening monitoring data, 2) AERMOD’s time-space mismatch, and
3) Proposed 50th % Bkg Method
Computational fluid dynamics in water waste treatment plantscm_teixeira
A CFD model can simulate the hydrodynamics of a design before implementation:
- Reducing lead-up times and costs
- Ultimately lead to optimization of reactor configuration
Summer Training 2015 at Alternate Hydro Energy CenterKhusro Kamaluddin
This is the presentation i gave to "Defend" my Summer Training at AHEC IIT Roorkee During Summer 2015. I gave this presentation in my college during my final year. Indeed the most lengthy i ever gave .
Determination of the corrosion rate of a mic influenced pipeline using four c...GeraldoRossoniSisqui
1) Gasunie used data from 4 pig runs over 5 years on a pipeline affected by MIC to determine the corrosion rate. 3 approaches were used: the first calculated a single rate for all defects, the second calculated individual rates, and the third used maximum likelihood estimation.
2) The third approach, which took into account measurement errors and ensured positive rates, estimated an average rate of 0.20 mm/yr with initiation largely at installation.
3) Analyzing shallow and deep defects separately found average rates of 0.23 mm/yr and 0.25 mm/yr respectively, indicating depth does not significantly impact the rate.
PetroSync - Advanced Seismic Data Acquisition and ProcessingPetroSync
This document provides information about an upcoming training course on advanced seismic data acquisition and processing. The 5-day course will cover key principles in data acquisition, survey design and execution, data preparation and pre-processing, advanced data processing techniques, and data interpretation. It will provide a holistic look at seismic acquisition and processing workflows and include case studies. The course aims to help participants implement optimal workflows, ensure quality control, and make business decisions regarding seismic surveys. It is intended for exploration geophysicists, interpreters, managers, and others involved in seismic data acquisition and processing.
ALTERNATE METHOD TO COMBINE MONITORED AND PREDICTED CONCENTRATIONS IN DISPERS...Sergio A. Guerra
The advent of the short term NAAQS has prompted us to reassess the level of conservatism commonly used in dispersion modeling analyses. An area of conservatism in NAAQS demonstrations relates to the combining of predicted (modeled) concentrations from AERMOD with observed (monitored) concentrations. Normally, some of the highest monitored observations are combined with AERMOD results yielding a very conservative total concentration. For example, in the case of the 1-hour NO2 NAAQS, the chances of the 98th percentile monitored concentration occurring at the same time as the meteorology to generate the 98th percentile ambient concentration is extremely low. Therefore, assuming that both of these happen at the same time is overly conservative.
This presentation includes justification for the use of a reasonable background concentration to combine with the AERMOD predicted concentration. The use of this method, if accepted by regulatory agencies, can help facilities demonstrate compliance in dispersion modeling analyses by assuming a more reasonable background concentration while, at the same time being protective of the NAAQS.
The Powerpoint presentation discusses about the Introduction to CFD and its Applications in various fields as an Introductory topic for Mechanical Engg. Students in General.
This document describes algorithms for detecting single radio pulses in real-time using graphics processing units (GPUs). It presents two new algorithms that use incomplete sets of boxcar filters to detect pulses at accelerated speeds with minimal signal loss. The algorithms were tested on simulated data and were found to process data 266-500 times faster than real-time on GPUs, detecting pulses with a mean 7% reduction in signal power.
Pairing aermod concentrations with the 50th percentile monitored valueSergio A. Guerra
Presentation delivered to the Background Concentrations Workgroup for Air Dispersion Modeling organized by the Minnesota Pollution Control Agency. delivered on March 25, 2014. Three topics covered include 1) Screening monitoring data, 2) AERMOD’s time-space mismatch, and
3) Proposed 50th % Bkg Method
Computational fluid dynamics in water waste treatment plantscm_teixeira
A CFD model can simulate the hydrodynamics of a design before implementation:
- Reducing lead-up times and costs
- Ultimately lead to optimization of reactor configuration
Summer Training 2015 at Alternate Hydro Energy CenterKhusro Kamaluddin
This is the presentation i gave to "Defend" my Summer Training at AHEC IIT Roorkee During Summer 2015. I gave this presentation in my college during my final year. Indeed the most lengthy i ever gave .
Determination of the corrosion rate of a mic influenced pipeline using four c...GeraldoRossoniSisqui
1) Gasunie used data from 4 pig runs over 5 years on a pipeline affected by MIC to determine the corrosion rate. 3 approaches were used: the first calculated a single rate for all defects, the second calculated individual rates, and the third used maximum likelihood estimation.
2) The third approach, which took into account measurement errors and ensured positive rates, estimated an average rate of 0.20 mm/yr with initiation largely at installation.
3) Analyzing shallow and deep defects separately found average rates of 0.23 mm/yr and 0.25 mm/yr respectively, indicating depth does not significantly impact the rate.
Computational fluid dynamics (CFD) is a tool for analyzing systems involving fluid flow, heat transfer and associated phenomena like chemical reactions using computer-based simulations. It involves numerically solving the governing equations of fluid flow to model the flow of liquids and gases. CFD complements experimental and theoretical fluid dynamics by providing a cost-effective means of simulating real flows. It has various applications in aerospace, automotive, turbo machinery, power plants, buildings, environmental engineering, and biomedical areas.
This document discusses computational fluid dynamics (CFD). CFD uses numerical analysis and algorithms to solve and analyze fluid flow problems. It can be used at various stages of engineering to study designs, develop products, optimize designs, troubleshoot issues, and aid redesign. CFD complements experimental testing by reducing costs and effort required for data acquisition. It involves discretizing the fluid domain, applying boundary conditions, solving equations for conservation of properties, and interpolating results. Turbulence models and discretization methods like finite volume are discussed. The CFD process involves pre-processing the problem, solving it, and post-processing the results.
techniques for detecting nanoparticles in wastewaterAlekhya Golla
techniques for seperating and analysing the metal containing nanoparticles in wastewater. these techniques are helpful in recycling process of industrial effluents containing nanowastes.
This document discusses recent trends in computational fluid dynamics (CFD). It begins by defining CFD as using numerical analysis and algorithms to solve fluid flow problems described by partial differential equations. CFD offers advantages over physical experiments by enabling low-cost simulation-based design and analysis of fluid phenomena that are difficult to measure experimentally. The document outlines the basic CFD process of geometry description, model selection, grid generation, solution, and post-processing. It provides examples of CFD applications in aerospace, automotive, biomedical, and other industrial fields to analyze designs. The conclusion discusses iterative solution methods and potential future advances in multidisciplinary and on-demand CFD simulations.
Computational fluid dynamics (CFD) is used to predict fluid flow, heat transfer, chemical reactions, and related phenomena. It solves governing equations for conservation of mass, momentum, energy, and species numerically. CFD analysis complements experimental testing and is used in conceptual design, product development, troubleshooting, and redesign across many industries. Example applications discussed in the document include missile engineering, architecture, food processing, automobiles, biomedical, turbomachinery, material transport chutes, and ocean current turbines. A variety of commercial CFD software packages are also listed.
This document reports on the development of an analytic model to calculate skyshine radiation doses from radioactive waste stores. Skyshine occurs when radiation reflects off the atmosphere back towards the ground. The analytic model is compared to results from Monte Carlo N-Particle (MCNP) simulations, which are considered the standard approach. The analytic model underestimated potential skyshine doses compared to MCNP, ranging from 43.4 to 188.1 times lower. This difference is mainly attributed to the analytic model not accurately accounting for scattering events in the concrete roof and air. The report concludes the analytic model is not suitable for radioactive waste store design without improvements to model scattering.
Computational fluid dynamics (CFD) is the use of computing to simulate fluid flow, heat transfer, and other related phenomena. CFD works by numerically solving the governing equations of fluid dynamics. It allows for analyzing flows that are difficult to study experimentally. CFD has various applications in fields like aerospace, automotive, biomedical, and power generation. The CFD process involves discretizing the domain, applying initial and boundary conditions, numerically solving the governing equations, and post-processing the results. Common discretization methods are finite volume, finite element, and finite difference methods. CFD provides insight into flows and heat transfer while being faster and cheaper than physical experiments.
This document provides an overview of computational fluid dynamics (CFD). It defines CFD as using computer simulations to predict fluid flow phenomena by modeling continuous fluids with partial differential equations. The document outlines where CFD is used in various industries like aerospace, automotive, biomedical, and more. It also discusses the physics, modeling, numerics, and overall CFD process involved in simulations. Examples are given of mesh generation and CFD being used to analyze problems like bottle filling.
This document compares computational fluid dynamics (CFD) modeling and physical flow modeling. Both methods can accurately simulate fluid flow behavior but have advantages and disadvantages. CFD models are usually faster and less expensive than physical models. However, physical models are better for simulating some phenomena like particulate behavior. For complex projects, using both CFD and physical modeling may provide the most comprehensive results. The appropriate modeling method depends on the specific application and objectives.
2 d and 3d land seismic data acquisition and seismic data processingAli Mahroug
The seismic method has three important/principal applications
a. Delineation of near-surface geology for engineering studies, and coal and mineral
exploration within a depth of up to 1km: the seismic method applied to the near –
surface studies is known as engineering seismology.
b. Hydrocarbon exploration and development within a depth of up to 10 km: seismic
method applied to the exploration and development of oil and gas fields is known
as exploration seismology.
c. Investigation of the earth’s crustal structure within a depth of up to 100 km: the
seismic method applies to the crustal and earthquake studies is known as
earthquake seismology.
Fluid dynamics, actually is the study of fluid under motion, governed with a certain set of conservation equations, wherein things are conserved, with reference to mass, momentum & energy.
If these three quantities i.e. mass, momentum & energy are solved entirely we can define any fluid flow. The conservation laws are formulated in the form of equations which we try to solve and that’s what simulation is all about. For my blogs kindly visit: https://www.learncax.com/knowledge-base/blog/by-author/ganesh-visavale
The document describes an optimization approach for underground sublevel stope mine planning that accounts for grade risk. It introduces the AMIRA Stope Optimizer software, which was used to optimize a project at Boliden's Älgträsk deposit. The optimization at a 1.8 g/t gold cut-off grade resulted in a 35% higher NPV than the break-even grade. The approach was extended to include grade risk quantification using block model simulations. By optimizing at different confidence levels, the project's economic risk can be assessed to aid decision making.
This document provides an introduction to computational fluid dynamics (CFD). It discusses the history of fluid dynamics from antiquity to the modern development of CFD. Key figures who contributed to the field are highlighted, including Archimedes, Leonardo da Vinci, Isaac Newton, Daniel Bernoulli, and Osborne Reynolds. The document also describes how CFD works by setting up the mathematical model, creating the mesh, solving the equations numerically, and examining the results. Applications of CFD and its advantages are discussed.
This document provides an introduction and overview of computational fluid dynamics (CFD) and the course "Introduction to Computational Fluid Dynamics". The key points covered are:
- The course covers numerical methods in CFD, including discretization techniques, accuracy, stability, grids, boundary conditions, and modeling turbulent flow.
- The learning objectives are for students to understand the role of computation in fluid dynamics and gain practical knowledge in setting up and analyzing simple aerodynamic problems using CFD.
- The course contents include introduction to partial differential equations, finite difference methods, grids, the Euler and Navier-Stokes equations, and case studies of best practices in CFD applications. Hands-on lab sessions make up
This presentation gives a brief introduction to the concept of coupled CFD-DEM Modeling.
Link to file: https://drive.google.com/open?id=1nO2n49BwhzBtT6NnvpxADG5WsC9uMJ-i
This document describes algorithms for quantitative precipitation estimation (QPE), 3D mosaic, and hydrometeor classification using polarimetric radar data in China. It details the radar network and scan strategies used, as well as algorithms for radar data quality control, remapping raw data to a Cartesian grid, and mosaicking reflectivity from multiple radars. It also describes a fuzzy logic method for hydrometeor classification based on polarimetric radar measurements, and plans for algorithm prototyping and validation.
The presentation gives glances of the importance of CFD analysis in engineering design with an illustration/ case study. For more details, follow the webinar "Role of CFD in Engineering Design" on https://www.learncax.com/knowledge-base/blog/by-author/ganesh-visavale
This document provides an overview of a Computational Fluid Dynamics (CFD) training course held from May 22-27, 2017 in Mumbai, India. The course will cover the basics of CFD including definitions, how CFD can help with design, the analysis process and steps, governing equations, input requirements, boundary and initial conditions, turbulence modeling, and numerical solution methods. The instructor has over 15 years of experience in hydraulic design engineering and will ensure attendees have a strong understanding of theoretical fluid dynamics and heat transfer needed to properly apply and interpret CFD simulations.
This document provides an overview of computational fluid dynamics (CFD) and the history of fluid dynamics. It discusses key figures from antiquity to the present who contributed to the development of fluid dynamics and CFD through experimental and theoretical work. These include Archimedes, Leonardo da Vinci, Isaac Newton, Osborne Reynolds, and Prandtl. The document also describes how CFD works by setting up mathematical models and discretizing domains into meshes before numerically solving the governing equations.
The presentation outlines the Adaptation Fund's Readiness Programme which aims to increase the number of accredited national implementing entities (NIEs) and strengthen NIE capacity. It describes the programme's activities including technical assistance grants, peer learning events, guidance documents, and workshops. Small grants are available in 2016 to support areas like environmental and social risk management, gender mainstreaming, and project formulation assistance. The goal is to help countries strengthen their ability to manage climate finance.
1) The document contains notes from a math lesson including assignments, warm-up problems, and examples of converting fractions to decimals and ordering decimals. It provides instructions to complete investigation problems 1-43 for homework due the next day and works through sample problems converting fractions to decimals and ordering decimals from least to greatest.
Computational fluid dynamics (CFD) is a tool for analyzing systems involving fluid flow, heat transfer and associated phenomena like chemical reactions using computer-based simulations. It involves numerically solving the governing equations of fluid flow to model the flow of liquids and gases. CFD complements experimental and theoretical fluid dynamics by providing a cost-effective means of simulating real flows. It has various applications in aerospace, automotive, turbo machinery, power plants, buildings, environmental engineering, and biomedical areas.
This document discusses computational fluid dynamics (CFD). CFD uses numerical analysis and algorithms to solve and analyze fluid flow problems. It can be used at various stages of engineering to study designs, develop products, optimize designs, troubleshoot issues, and aid redesign. CFD complements experimental testing by reducing costs and effort required for data acquisition. It involves discretizing the fluid domain, applying boundary conditions, solving equations for conservation of properties, and interpolating results. Turbulence models and discretization methods like finite volume are discussed. The CFD process involves pre-processing the problem, solving it, and post-processing the results.
techniques for detecting nanoparticles in wastewaterAlekhya Golla
techniques for seperating and analysing the metal containing nanoparticles in wastewater. these techniques are helpful in recycling process of industrial effluents containing nanowastes.
This document discusses recent trends in computational fluid dynamics (CFD). It begins by defining CFD as using numerical analysis and algorithms to solve fluid flow problems described by partial differential equations. CFD offers advantages over physical experiments by enabling low-cost simulation-based design and analysis of fluid phenomena that are difficult to measure experimentally. The document outlines the basic CFD process of geometry description, model selection, grid generation, solution, and post-processing. It provides examples of CFD applications in aerospace, automotive, biomedical, and other industrial fields to analyze designs. The conclusion discusses iterative solution methods and potential future advances in multidisciplinary and on-demand CFD simulations.
Computational fluid dynamics (CFD) is used to predict fluid flow, heat transfer, chemical reactions, and related phenomena. It solves governing equations for conservation of mass, momentum, energy, and species numerically. CFD analysis complements experimental testing and is used in conceptual design, product development, troubleshooting, and redesign across many industries. Example applications discussed in the document include missile engineering, architecture, food processing, automobiles, biomedical, turbomachinery, material transport chutes, and ocean current turbines. A variety of commercial CFD software packages are also listed.
This document reports on the development of an analytic model to calculate skyshine radiation doses from radioactive waste stores. Skyshine occurs when radiation reflects off the atmosphere back towards the ground. The analytic model is compared to results from Monte Carlo N-Particle (MCNP) simulations, which are considered the standard approach. The analytic model underestimated potential skyshine doses compared to MCNP, ranging from 43.4 to 188.1 times lower. This difference is mainly attributed to the analytic model not accurately accounting for scattering events in the concrete roof and air. The report concludes the analytic model is not suitable for radioactive waste store design without improvements to model scattering.
Computational fluid dynamics (CFD) is the use of computing to simulate fluid flow, heat transfer, and other related phenomena. CFD works by numerically solving the governing equations of fluid dynamics. It allows for analyzing flows that are difficult to study experimentally. CFD has various applications in fields like aerospace, automotive, biomedical, and power generation. The CFD process involves discretizing the domain, applying initial and boundary conditions, numerically solving the governing equations, and post-processing the results. Common discretization methods are finite volume, finite element, and finite difference methods. CFD provides insight into flows and heat transfer while being faster and cheaper than physical experiments.
This document provides an overview of computational fluid dynamics (CFD). It defines CFD as using computer simulations to predict fluid flow phenomena by modeling continuous fluids with partial differential equations. The document outlines where CFD is used in various industries like aerospace, automotive, biomedical, and more. It also discusses the physics, modeling, numerics, and overall CFD process involved in simulations. Examples are given of mesh generation and CFD being used to analyze problems like bottle filling.
This document compares computational fluid dynamics (CFD) modeling and physical flow modeling. Both methods can accurately simulate fluid flow behavior but have advantages and disadvantages. CFD models are usually faster and less expensive than physical models. However, physical models are better for simulating some phenomena like particulate behavior. For complex projects, using both CFD and physical modeling may provide the most comprehensive results. The appropriate modeling method depends on the specific application and objectives.
2 d and 3d land seismic data acquisition and seismic data processingAli Mahroug
The seismic method has three important/principal applications
a. Delineation of near-surface geology for engineering studies, and coal and mineral
exploration within a depth of up to 1km: the seismic method applied to the near –
surface studies is known as engineering seismology.
b. Hydrocarbon exploration and development within a depth of up to 10 km: seismic
method applied to the exploration and development of oil and gas fields is known
as exploration seismology.
c. Investigation of the earth’s crustal structure within a depth of up to 100 km: the
seismic method applies to the crustal and earthquake studies is known as
earthquake seismology.
Fluid dynamics, actually is the study of fluid under motion, governed with a certain set of conservation equations, wherein things are conserved, with reference to mass, momentum & energy.
If these three quantities i.e. mass, momentum & energy are solved entirely we can define any fluid flow. The conservation laws are formulated in the form of equations which we try to solve and that’s what simulation is all about. For my blogs kindly visit: https://www.learncax.com/knowledge-base/blog/by-author/ganesh-visavale
The document describes an optimization approach for underground sublevel stope mine planning that accounts for grade risk. It introduces the AMIRA Stope Optimizer software, which was used to optimize a project at Boliden's Älgträsk deposit. The optimization at a 1.8 g/t gold cut-off grade resulted in a 35% higher NPV than the break-even grade. The approach was extended to include grade risk quantification using block model simulations. By optimizing at different confidence levels, the project's economic risk can be assessed to aid decision making.
This document provides an introduction to computational fluid dynamics (CFD). It discusses the history of fluid dynamics from antiquity to the modern development of CFD. Key figures who contributed to the field are highlighted, including Archimedes, Leonardo da Vinci, Isaac Newton, Daniel Bernoulli, and Osborne Reynolds. The document also describes how CFD works by setting up the mathematical model, creating the mesh, solving the equations numerically, and examining the results. Applications of CFD and its advantages are discussed.
This document provides an introduction and overview of computational fluid dynamics (CFD) and the course "Introduction to Computational Fluid Dynamics". The key points covered are:
- The course covers numerical methods in CFD, including discretization techniques, accuracy, stability, grids, boundary conditions, and modeling turbulent flow.
- The learning objectives are for students to understand the role of computation in fluid dynamics and gain practical knowledge in setting up and analyzing simple aerodynamic problems using CFD.
- The course contents include introduction to partial differential equations, finite difference methods, grids, the Euler and Navier-Stokes equations, and case studies of best practices in CFD applications. Hands-on lab sessions make up
This presentation gives a brief introduction to the concept of coupled CFD-DEM Modeling.
Link to file: https://drive.google.com/open?id=1nO2n49BwhzBtT6NnvpxADG5WsC9uMJ-i
This document describes algorithms for quantitative precipitation estimation (QPE), 3D mosaic, and hydrometeor classification using polarimetric radar data in China. It details the radar network and scan strategies used, as well as algorithms for radar data quality control, remapping raw data to a Cartesian grid, and mosaicking reflectivity from multiple radars. It also describes a fuzzy logic method for hydrometeor classification based on polarimetric radar measurements, and plans for algorithm prototyping and validation.
The presentation gives glances of the importance of CFD analysis in engineering design with an illustration/ case study. For more details, follow the webinar "Role of CFD in Engineering Design" on https://www.learncax.com/knowledge-base/blog/by-author/ganesh-visavale
This document provides an overview of a Computational Fluid Dynamics (CFD) training course held from May 22-27, 2017 in Mumbai, India. The course will cover the basics of CFD including definitions, how CFD can help with design, the analysis process and steps, governing equations, input requirements, boundary and initial conditions, turbulence modeling, and numerical solution methods. The instructor has over 15 years of experience in hydraulic design engineering and will ensure attendees have a strong understanding of theoretical fluid dynamics and heat transfer needed to properly apply and interpret CFD simulations.
This document provides an overview of computational fluid dynamics (CFD) and the history of fluid dynamics. It discusses key figures from antiquity to the present who contributed to the development of fluid dynamics and CFD through experimental and theoretical work. These include Archimedes, Leonardo da Vinci, Isaac Newton, Osborne Reynolds, and Prandtl. The document also describes how CFD works by setting up mathematical models and discretizing domains into meshes before numerically solving the governing equations.
The presentation outlines the Adaptation Fund's Readiness Programme which aims to increase the number of accredited national implementing entities (NIEs) and strengthen NIE capacity. It describes the programme's activities including technical assistance grants, peer learning events, guidance documents, and workshops. Small grants are available in 2016 to support areas like environmental and social risk management, gender mainstreaming, and project formulation assistance. The goal is to help countries strengthen their ability to manage climate finance.
1) The document contains notes from a math lesson including assignments, warm-up problems, and examples of converting fractions to decimals and ordering decimals. It provides instructions to complete investigation problems 1-43 for homework due the next day and works through sample problems converting fractions to decimals and ordering decimals from least to greatest.
This document discusses prepositions and provides examples of their uses. Prepositions show place, position, time, direction, and method. Pictures are used to demonstrate sentences with prepositions like "under", "on", "behind", and "in front of". Learners are given activities to practice identifying prepositions in sentences and constructing their own sentences using prepositions like "over", "down", "round", "from", and "up". The lesson on prepositions is presented by Waneroba D. Watuha from Mivumoni Secondary School in Kwale County.
The document contains notes from a math lesson on October 14, 2013. It lists assignments due, including a test on proportions. The warm-up questions cover converting a percentage to a fraction, finding a percentage of a number, listing perfect squares, and simplifying expressions. The lesson introduces proportions and shows examples of setting up and solving proportions using cross multiplication.
Mr. Anderson teaches a character development class that focuses on respect. Respect is defined as showing high regard for oneself, others, and property. The class discusses treating others the way you want to be treated through examples like asking permission before using others' things and showing respect for people who are different. The class covers quotes about respect and related concepts like courtesy, tolerance, and appreciation. Students read a fable about a boy who gets his hand stuck in a jar while grabbing too many nuts and refuses to let any go. They are asked to consider what the moral of the story means and whether it is better to have something divided evenly or get more than one's share.
As a part of the 30% exploratory IP/ tech based products/ projects we used to do in And Then. We decided to explore simplification and automation of data visualisations. So, corporate employees with zero design training and skills could generate and use designed infographics. Based on pre-set templates of colour and iconography. This project was a collaboration between me and one of our brightest design interns 'Santhosh'. It was a fabulous learning for me, and this documentation by Santhosh is a good summation.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive function. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
This PowerPoint presentation introduces subject-verb agreement and explores some of the difficulties in getting subjects and verbs to agree properly. It provides examples of singular and plural subjects requiring singular or plural verbs, respectively. It also examines tricky cases involving pronouns, fractions, compound subjects, and sentences beginning with "there." The presentation aims to help writers avoid agreement errors in their own writing.
Social advergaming | Joan Miquel Martinez | AdverwaySmash Tech
El documento describe los social advergames, juegos en redes sociales que promocionan productos o ideas de forma entretenida. Explica que los advergames deben durar lo suficiente para promocionar la marca, generar empatía con los jugadores de forma voluntaria y con publicidad poco intrusiva. También proporciona ejemplos exitosos como el juego de Monopoly de McDonald's y America's Army que generan exposición a la marca y ventas con pocos costes relacionados con la publicidad.
Soundbuzz is a digital music service provider focused on the Asia-Pacific market. It offers secure digital downloads of music, ringtones and images through online and wireless partnerships. Soundbuzz operates in the music, IT and marketing industries and provides a one-stop-shop for digital music commerce through technology, creative, editorial and legal services. It uses various e-commerce models including self-service, collaboration, information aggregation and extended enterprise patterns to allow customers to search for music, make purchases and access purchase history. The presentation evaluates Soundbuzz's business patterns and provides recommendations on how to improve its digital music offerings.
A Vedic Maths is the name given to the ancient system of Indian Mathematics which was rediscovered from the Vedas/sutras between 1911 and 1918 by Sri Bharati Krisna Tirthaji (1884-1960).
According to his research, maths is based on 16 SUTRAS or word-formulae. These formulae describe the way the mind works naturally and are therefore a great help in directing the students to the appropriate solution. This unifying quality is very satisfying,; it makes maths easy, enjoyable and encourages innovation.
Sammaan is a design studio that aims to economically empower impoverished women through upcycling discarded furniture. Their first product was "Second Life" chairs reconstructed from discarded chairs by mothers participating in the LettuceBee Kids program. They have since expanded to developing calligraphy-inspired products using urdu poetry to revive the art form and heritage. The business model involves identifying low-income communities with relevant skills, collaborating with partners for materials, production, and distribution, and marketing unique high-quality pieces to socially-conscious consumers through various sales channels.
How do we translate good ideas into great mobile apps? We follow a seven step process. From idea to engineering our process results in mobile solutions that delight the consumer. Emotion Engineering & Design at play.
Live Q&A Session with Kashyap Vadapalli,PepperfryUpGrad
This document outlines the agenda for a live Q&A session with Kashyap Vadapalli, CMO of Pepperfry, including discussions around branding, how Pepperfry developed their brand, measuring customer feedback through Net Promoter Score, and using trial pricing for new products. The branding example of Pepperfry will be provided, which established the brand promise of making the anxious furniture buying process a happy experience through standardization and providing information.
The document provides an operations manual for HP Hotels that outlines various policies and procedures for the company. It begins with an introduction stating that the manual establishes minimum guidelines for standardized and cost-efficient hotel operations. It then provides a mission statement focusing on profitable growth and shareholder value. The manual continues with sections covering various operational areas like accounting, administration, food and beverage, front desk, housekeeping, maintenance, sales, and human resources. It aims to standardize guidelines across hotel properties while allowing for good judgment in decision making.
This document is a thesis submitted by Tokelo Khalema to the University of the Free State in partial fulfillment of the requirements for a B.Sc. Honors degree in Mathematical Statistics. The thesis compares the Gaussian linear model, two Bayesian Student-t regression models, and the method of least absolute deviations through a Monte Carlo simulation study. The study aims to evaluate how soon and how severely the least squares regression model starts to lose optimality against these robust alternatives under violations of its assumptions. The document includes sections on robust statistical procedures, literature review of the models considered, research methodology, results and applications of the simulation study, and closing remarks.
This thesis aims to develop deep learning models to detect COVID-19 pneumonia in chest X-ray images. The author trains two models: 1) A binary classifier to distinguish COVID-19 pneumonia from non-COVID cases, which classifies all test cases correctly. 2) A four-class classifier to identify COVID-19, viral pneumonia, bacterial pneumonia, and normal cases, which achieves an average accuracy of 93% on the test set. Gradient-weighted Class Activation Mapping is used to interpret the four-class model and finds it can focus on patchy areas characteristic of COVID-19 pneumonia to make accurate predictions.
This document is a dissertation submitted by Chunmei Shi to Stanford University in partial fulfillment of the requirements for a Doctor of Philosophy degree. The dissertation investigates the flow behavior of gas-condensate wells through experimental and numerical modeling. Chapter 1 introduces the problem of productivity losses in gas-condensate reservoirs due to liquid dropout near the wellbore. Chapters 2-5 describe experimental investigations of gas-condensate flow through a core and numerical simulations to model compositional variations. Chapter 5 also uses optimization techniques to improve recovery. The conclusions indicate the work developed theoretical models, experimentally studied flow in a core, and numerically simulated gas-condensate flow to increase productivity from gas-condensate reservoirs.
This dissertation proposes and analyzes methods for adaptive transceiver design and performance enhancement in OFDM systems. It first develops a new model for the time-varying indoor wireless channel based on Doppler spectrum measurements. It then presents adaptive synchronization techniques for OFDM receivers, including windowing and Kalman filtering approaches, to mitigate timing estimation errors under different channel conditions. An enhanced DFT-based MMSE channel estimator using Kalman smoothing is also proposed to minimize spectral leakage effects from virtual carriers. Finally, a method is introduced to improve throughput by adaptively optimizing constellation sizes and power distributions across sub-carriers during transmissions. Simulation results show the proposed estimation methods and throughput optimization can significantly improve synchronization accuracy and system performance.
This thesis studies the design optimization and modeling improvement of a point absorber wave energy converter. It evaluates three hull shapes (cylinder, bullet, cone) through frequency and time domain models to determine the most efficient design. It also develops two final models: Model 1 includes viscous drag forces more accurately, finding it can significantly decrease predicted power extraction. Model 2 assesses the impact of varying wave force position on the hull, but finds it provides little benefit computationally. The thesis aims to advance point absorber design and modeling.
This document presents research on developing an emissions prediction capability for use in conceptual aircraft engine design. A physics-based modeling approach was used consisting of 1D flow models of different combustor types, chemical reactor networks representing the combustor flowfields, and regression models. Models were developed for a single annular combustor, rich-quench-lean combustor, and lean-burn combustor. The models were implemented in NPSS and validated against test data. Regression models were also created to rapidly predict emissions for parametric studies of a CFM56-sized engine application. The new capability allows emissions predictions during early engine conceptual design when cycle parameters and architectures are varied.
This document is the thesis presented by Joanie Michellene Claudette Geldenhuys to Stellenbosch University for the degree of Master of Science in Electrical and Electronic Engineering. The thesis investigates the use of model predictive control for current control of a three-phase grid-tied voltage source converter with an LCL filter. A cost function is formulated to minimize reference tracking error and switching frequency. The grid voltage is incorporated into the system model as an additional input. Simulation results show the controller provides fast transient response and good reference tracking at high switching frequencies but is unable to meet harmonic limits at low switching frequencies as required by the South African grid code.
Breath analysis by quantum cascade spectroscopy - Master thesis by Olav Grouw...Olav Grouwstra
This thesis examines breath analysis using quantum cascade laser spectroscopy to distinguish between healthy and asthmatic patients. It describes the setup used, which includes a quantum cascade laser, multipass cavity, distance meter, and mercury cadmium telluride detectors. The document details the data calibration process to reduce wavenumber uncertainties and calibrate to CO2 and H2O measurements. It then determines the molecules present, their concentrations, and removes CO2 to estimate ethanol levels. Finally, it applies machine learning algorithms to classify samples by health status and predict diseases from their molecular compositions and concentrations.
This document is a thesis submitted by Yishi Lee to Embry-Riddle Aeronautical University for the degree of Master of Science in Engineering Physics. The thesis developed a new 3-D model of the high-latitude ionosphere to study the coupling between the ionosphere, magnetosphere, and neutral atmosphere. The model consists of equations describing the conservation of mass, momentum, and energy for six ionospheric constituents, as well as an electrostatic potential equation. The thesis was prepared under the direction of Dr. Matthew Zettergren and other committee members. It uses the 3-D model to examine ion heating, plasma structuring due to perpendicular transport, ion upflow, molecular ion generation, and neutral wave forcing in
This master's thesis develops a mathematical index called the Γ-index to quantify the technological flexibility of dispatchable power generation units. The Γ-index is calculated as the normalized integral of weighted flexibility terms over time. These terms evaluate dynamic features like ramp rates and minimum run times, as well as static features such as maximum step changes and reliability. The Γ-index is tested by evaluating the flexibility of different power plant technologies. The results show diesel engines and gas turbines are most flexible, followed by combined cycle plants, then steam plants, with nuclear ranked as least flexible. The Γ-index could help quantify the value of flexible generation and inform decisions around power system planning and policymaking.
This document presents the thesis of Eduardo Wilde Barbaro on the interaction between aerosols and convective boundary-layer dynamics over land. The thesis committee is listed, consisting of professors and researchers from Wageningen University and other institutions. The thesis will be defended on March 20, 2015 in a public presentation to discuss Barbaro's research on modeling the coupling between aerosols, radiation, thermodynamics and land surface processes in the atmospheric boundary layer.
The document is a project report on parametric design and optimization of propellers using Grasshopper. It discusses using Grasshopper to build a parametric model of the Wageningen B-screw propeller series. The report reviews literature on ship propellers and propulsion. It then analyzes the geometry and performance characteristics of the Wageningen B-screw series. The main body of the report describes realizing a parametric model of the propeller series in Grasshopper that allows modifying the geometry and calculating performance based on parameters. The model is then tested and potential further work is discussed.
This dissertation examines modeling considerations for long-term generation and transmission expansion planning to address increasing load uncertainty and renewable generation. The author develops an optimization framework using an outer optimization to search for candidate upgrades and an inner optimization to evaluate cost and reliability. Simulations are performed on the IEEE 14 bus system comparing a unified generation and transmission expansion planning model to a sequential one. Sensitivity analyses evaluate the impact of perturbations and load uncertainty on system cost and reliability. The results provide insights into planning with high renewable penetration under uncertainty.
Performance Evaluation of Path Planning Techniques for Unmanned Aerial VehiclesApuroop Paleti
This thesis evaluates the performance of the A-star algorithm and Mixed Integer Linear Programming (MILP) for path planning of unmanned aerial vehicles (UAVs). The author conducted simulation experiments to compare the computational time and efficiency of the two methods in different scenarios. The key findings are:
1) MILP had significantly lower computation times than A-star for both scenarios, taking on average 10-12 seconds per function compared to 1-3 seconds for A-star.
2) Statistical analysis via hypothesis testing found a significant difference between the computation times of the two algorithms.
3) Based on the results, MILP was found to be more efficient than A-star for path planning
This document presents an honours thesis that aims to develop a parametric approach to analyzing labour productivity growth. The thesis relaxes key assumptions of the standard growth accounting framework, including allowing for variable returns to scale, technical inefficiency, and statistical noise. A stochastic frontier analysis is used to estimate a production function for multiple countries over 30 years. Labour productivity growth is then decomposed into contributions from technological change, input deepening, efficiency changes, scale effects, and noise. The decomposition is applied to 15 countries' agricultural sectors from 1977 to 2007.
Leakage detection in water pipe networks using Ground Penetrating Radar (GPR)...Dai Shi
This document presents a master's thesis on using ground penetrating radar (GPR) for detecting water leaks in pipe networks. The thesis includes numerical simulations to evaluate GPR's sensitivity to different parameters and leak configurations. Laboratory experiments were conducted using a sandbox to monitor controlled leaks. A field measurement was also taken of a pre-located leak in an urban area. The numerical simulations and experiments provided insights into GPR's performance and limitations for leak detection. The thesis evaluates GPR's potential for precisely locating leaks and imaging underground structures compared to traditional acoustic methods.
This MSc thesis examines the stable isotopic composition of atmospheric water vapor in Copenhagen, Denmark. Two Picarro cavity ring-down spectrometers were used to continuously sample vapor over a 24-hour period to obtain high temporal resolution isotopic data. Meteoric water samples were also collected using a rain collector. Isotopic data from the vapor phase were combined with precipitation samples to interpret the link between vapor and liquid/solid water isotopic compositions. The results provide a better understanding of how the two spectrometers perform for vapor sampling and allow for comparison of the isotopic time series data.
This document is a dissertation on pedestrian traffic simulation and experiments. It presents a discrete model of pedestrian motion and evaluates it against empirical data from evacuation exercises and experiments. The model is able to accurately reproduce fundamental diagrams from empirical studies and is computationally efficient for large-scale simulations.
1. Kalman filter predictor
applied to air quality forecasts
Bachelor’s Thesis
Oskar Triebe
June 20, 2015
Supervisor at ETH: Prof. Dr. Jing Wang
Supervisor at HKUST: Prof. Dr. Alexis Lau
Advisors at HKUST: Dr. Ying Li, Mr. Tong Liu
Department of Civil, Environmental and Geomatic Engineering, ETH Z¨urich
2.
3. Abstract
The Kalman filter (KF) is a recursive post processing method gaining
popularity in the field of air quality (AQ) forecasting. It is can be used
to remove bias from model predictions. Preceding studies achieved
this by modelling the difference of predictions and observations (D-
KF). This study aims to provide a larger set of methods to correct AQ
forecasts. Two alternative bias modelling techniques are additionally
introduced: The logarithmic ratio bias (LR-KF) and the relative differ-
ence bias (RD-KF). Complimented with further variations, these were
applied on the forecasts with the aim to remove the model bias and
improve the prediction of high pollution episodes (HPE). Three years
of hourly concentrations of fine suspended particulates (PM25), nitro-
gen dioxide (NO2) and ozone (O3) at three stations in Hong Kong SAR
were utilised. An optimal method for each station and pollutant was
selected by evaluating the normalised mean square error (NMSE), un-
paired peak prediction accuracy (UPPA), fractional bias (FB), fraction
of predictions within a factor of two (FAC2) and a measure of the skill
to predict HPE (CFM). A comparison with persistence, the trivial time-
lagging of observed concentrations by one day, showed that the raw
model forecasts perform significantly worse than this trivial method.
The selected methods showed a significantly improved overall perfor-
mance compared to both, the raw forecasts and persistence. The CFM
however was not improved in all cases. The optimal method was dif-
ferent for each pollutant and station with the exception of NO2 where
the RD-KF was most skilful. It was also found that the LR-KF is best at
correcting extreme differences and peaks, whereas the D-KF and RD-
KF perform better in general cases. Further, strengths and weaknesses
of the KF in general and of the different bias modelling techniques are
explained.
i
4.
5. Acknowledgements
I would like to acknowledge Professor Alexis Lau for his great guidance
and the valuable insights he offered me during our frequent meetings at the
Hong Kong University of Science and Technology, Doctor Ying Li and Tong
Liu for advising me in the process of creating this study. Also I would like
to thank David Yeung for his help to access the data. Finally I am grateful
for the support provided by Professor Jing Wang as our supervisor at our
home institution, ETH Zurich.
8. Contents
B Variance Computation and Initial Values 25
C Literature Review 27
D Statistical Meaures 31
D.1 Normalisation of Statistics and Selection of Best Methods . . 31
D.2 Normalised Mean Scores of All Methods and Models . . . . . 32
D.3 Statistical Measures for Corrected Model Ensemble Forecasts 33
Bibliography 43
vi
9. List of Tables
2.1 Ratio limits for LR and RD bias. The values are the maximum
ratio of Pi/Oi and Oi/Pi for each origin respectively, restricting
less than 1% of the upper and lower data points. . . . . . . . . . . 9
3.1 Performance of best methods for PM25 compared with model
performance and persistence. . . . . . . . . . . . . . . . . . . . . 12
3.2 Performance of best methods for NO2 compared with model per-
formance and persistence. . . . . . . . . . . . . . . . . . . . . . . 12
3.3 Performance of best methods for O3 compared with model per-
formance and persistence. . . . . . . . . . . . . . . . . . . . . . . 13
D.1 Normalised average scores of KF corrected forecasts for each
model. Mean taken over all methods. . . . . . . . . . . . . . . . . 32
D.2 Normalised average corrected forecast scores for each method.
Mean taken over all models. . . . . . . . . . . . . . . . . . . . . . 33
D.3 Statistical measures of corrected model ensemble forecasts for
PM25 at TW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
D.4 Statistical measures of corrected model ensemble forecasts for
PM25 at CL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
D.5 Statistical measures of corrected model ensemble forecasts for
PM25 at CB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
D.6 Statistical measures of corrected model ensemble forecasts for
NO2 at TW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
D.7 Statistical measures of corrected model ensemble forecasts for
NO2 at CL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
D.8 Statistical measures of corrected model ensemble forecasts for
NO2 at CB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
D.9 Statistical measures of corrected model ensemble forecasts for O3
at TW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
vii
10. List of Tables
D.10 Statistical measures of corrected model ensemble forecasts for O3
at CL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
D.11 Statistical measures of corrected model ensemble forecasts for O3
at CB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
viii
11. List of Figures
3.1 Hourly sample plot of PM25 at TW for the three different bias
modelling procedures. . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.2 Hourly sample plot of PM25 at TW for the four variants of RD. . 14
3.3 False model spike corrected by D, LR and RD. . . . . . . . . . . . 15
3.4 Peaks of observed and predicted concentrations modelled by D,
LR and RD. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.5 Extreme model bias changes modelled by D, LR and RD. . . . . . 17
C.1 Flow chart describing the Kalman filter predictor process, adapted
from Delle Monache et al.[2] . . . . . . . . . . . . . . . . . . . . . 29
ix
12.
13. Chapter 1
Introduction
Background and Motivation Air Quality is not a new concern, but more
and more people live in highly populated areas and become more affected
by the locally and regionally produced pollution. Ever since London’s great
smog of 1952, the science of air quality modelling has steadily developed.
Today these models are integrated with advanced numerical weather pre-
diction (NWP) models. Model output statistics (MOS) have also been exten-
sively researched. Traditional MOS such as different regression techniques
are mature and are a common post processing procedure for many models.
New methods such as neural networks have recently been developed and
are growing more mature. Although the Kalman filter algorithm was devel-
oped half a century ago, it has only been increasingly applied in the field of
air quality forecasting during the last decade. Some further insights can be
found in the appended literature review.
The AQ model executed for the Hong Kong SAR at the Hong Kong Univer-
sity of Science and Technology is not post processed by any MOS. The AQ
forecasts are currently affected by systematic and random errors.
The KF has also been successfully used to post process AQM forecasts and
improve the accuracy. However, a study done by Delle Monacea et al. [2]
showed that the KF decreased the forecast performance for some stations.
This suggests that the performance is dependent on the characteristics of
the model-observations relationships. All reviewed studies used the KF to
model the simple difference of observations and predictions as the model
bias. As the same modelled pollutant might show different patterns at each
analysed station, a different way of modelling the bias might be optimal for
different stations.
1
14. 1. Introduction
Scope This study aims to develop and evaluate different ways to model
and minimise the AQM forecast bias using the Kalman filter algorithm.
Other types of bias modelling such as the natural logarithm of the ratio
of forecasts versus observations or the relative difference bias will be intro-
duced. Further variations of the basic KF algorithm such as a weighted
average or different ways of modelling the variances will be examined.
In order to ensure the significance of the results, a long time period of multi-
ple pollutants and stations will be processed. The results will we evaluated
by statistical measures significant for the performance of AQM forecasts.
Objectives
1. Create different procedures utilising the KF algorithm to post process
AQM forecasts.
2. Assess the performance of the corrected forecasts and display weak-
nesses and strengths.
3. Suggest a customised procedure for each pollutant at each station.
2
15. Chapter 2
Methods
2.1 Kalman Filter Predictor
The whole Kalman filter algorithm is recursively based on the last estimated
state and its variances. In predictor mode, the latest state estimate is used
as a prediction of the next state. Jolliffe and Stephenson provide a good
description of the Kalman filter predictor (KFP): “The KFP correction is an
automatic post-processing method that uses the recent past observations
and forecasts to estimate the model bias in the future forecast, where bias
here is defined as the difference of the central location of the forecasts and
the observations”[3, 2].
In this study other methods of modelling the bias are also introduced in the
procedure section 2.4.1.
The equations of the KFP can be derived from the general KF equations.
A short introduction to the Kalman filter can be found in section A of the
appendices. In this case only one state parameter xt is modelled and no ad-
ditional system parameter µt are used. We define the signal input zt as yt to
make clear that it represents the bias of the AQM and not the concentrations
of the pollutants themselves. To further simplify the system, the matrices A,
B and C are chosen to be identity.
Now the simplified equations for the estimated (true) model bias xt can be
written as
ˆxt = xt−1 + qt−1
where qt is a normal distributed random error with zero mean and variance
σ2
q . The observed model bias yt is expected to differ from the estimated bias
by another normal distributed random error et with zero mean and variance
σ2
e .
ˆyt = ˆxt + et = xt−1 + qt−1 + et
3
16. 2. Methods
The state covariance matrix, or in this case the expected mean square error
of the state ˆRt is estimated to be
ˆRt = Rt−1 + σ2
q
Next, the Kalman gain Kt can be computed as
Kt =
ˆRt
ˆRt + σ2
e
=
Rt−1 + σ2
q
Rt−1 + σ2
q + σ2
e
Finally, the state xt and its covariance matrix Rt are updated
xt = ˆxt + Kt(yt − ˆyt) = xt−1 + Kt(yt − xt−1)
Rt = (1 − Kt) ˆRt = (1 − Kt)(Rt−1 + σ2
q )
The state estimate xt is now assumed to be the bias of the AQM forecast for
the next time step. When predicting, we can only use present observations to
correct the future prediction as the future observations are not yet available.
This is why this use of the KF is usually named KFP.
After every iteration, the whole system is time-shifted t = t + 1. In this study
a time step of one day is applied because the AQM forecasts are generated
once a day. For each day 24 independent KFP are run for each forecast hour.
The details about the computation of the variances and initial values can be
found in appendix B.
2.2 Data
The data of three full years, 2012-2014 at three locations, Tsuen Wan (TW),
Central / Western (CL) and Causeway Bay (CB) was used. Three pollutants,
fine suspended particulates of size smaller than 2.5 µm (PM25), nitrogen
dioxide (NO2) and ozone (O3), are analysed at each station.
Tsuen Wan is an ambient measuring station often used to represent Hong
Kong as it is located in the heart of the SAR HK. Central / Western is also an
ambient station in an area where many people work and live. As it is locked
in by the surrounding skyscrapers, it follows different dynamics than the
ambient station TW. Causeway Bay is a roadside station, also situated in a
densely populated area. As the station is located right next to a main road,
the patterns of AQ parameters is quite different from TW and CL. At CB the
AQM forecasts differ at times quite significantly from the observations.
The Hong Kong air quality health index (AQHI) is defined by PM25, NO2,
O3 and two further pollutants, respirable suspended particulates smaller
4
17. 2.2. Data
than 10 µm (PM10) and sulphur dioxide (SO2). As PM25 and PM10 are
strongly correlated and the control of SO2 in Hong Kong has improved, this
study is limited to PM25, NO2 and O3. These are also the three parameters
defining the Canadian AQHI.
Each station and pollutant combination will be treated as a separate data set
and will referred to from now on as a (data) origin. The statistical measures
for the raw data will be provided in the Results chapter when contrasted
with the best Kalman filtering methods. To avoid negative concentrations
and division by zero, a minimal threshold of 1.0 µgm−3 respectively 1.0 ppb
was set for all observations and model forecasts.
2.2.1 Observations
The reported concentration for each full hour is computed by the average
concentration of the last 60 minutes at that station. Missing observations
were replaced with the last reported concentration. In total, all stations
had less than 4% of missing values with the exceptions of CB, where 12.0%
of PM25 and 5.4% of O3 measurements were missing. All data sets meet
the requirement of less than 15% missing data in order to be considered
consistent.
2.2.2 Model Predictions
Air Quality forecasts computed by the U.S. EPA’s Community Multi-scale
Air Quality Model (CMAQ) and by the open-source Comprehensive Air
Quality Model with Extensions (CAMx) are evaluated. Both models are well
documented and will not be further explained in this study. The hourly con-
centrations generated by CMAQ are instant concentrations at the respective
hour, while CAMx delivers the average concentration of the past 60 minutes.
The models deliver a forecast for several future days. Theoretically, the fore-
cast of the next day would be optimal to use. However, the models usually
perform bad during their spin-up time of around 12 hours and due to the
time it takes for the model to run, the first forecast hours are already past
until the results are available. Due to the spin-up issues and the temporal
delays of delivering the results, the past days model forecast for the second
day is generally used as a prediction for the next day. Therefore this study
uses the previous days model second day forecast and post processes this
with the observations of the present day.
The AQM forecast data was complete with the exception of the first 36 hours
of CAMx in 2012. These missing values were replaced with the respective
CMAQ forecasts.
5
18. 2. Methods
Additional Models More model data was simulated by applying a weighted
average of the two models, taking the mean of the two models and of the av-
eraged models. The model ensemble forecasts are marked by an ’M-E’. Fur-
ther model data sets were obtained by combining already corrected forecasts
into a new data set. Hereby the corrected forecasts of the two original Model
inputs were averaged. This was done for each bias modelling method with
the basic KFP, the KFP with modelled variance and the KFP with modelled
variance and interconnected hours. The KFP variations and the smooth-
ing are explained in section 2.4.2. These Kalman filtered ensemble forecast
model inputs are labelled as ’Kf-E’. Lastly three means of all variations and
all bias types was taken for not smoothed model input, smoothed before and
one smoothed before and after. These data sets are marked by ’Kf-E-mean’.
This procedure lead to 18 model data sets in total.
2.3 Verification Statistics
The following statistical measures are used to evaluate the performance of
the corrected forecasts and to compare them to the corresponding raw fore-
casts. Five measures were selected to account for a balanced estimate of how
well the predicted concentrations (Pi) match the observations (Oi) in general
and at peak times. Hereby the first 10 days of 2012 are not evaluated as this
period is considered as a spin-up phase for the KF. The details about the
normalisation of the statistical measures and their combination into scores
can be found in the appendix D.2.
Normalised Mean Square Error The normalised mean square error (NMSE)
measures the overall deviation of predictions towards observations.
NMSE =
1
n
n
∑
i=1
(Oi − Pi)2
¯O · ¯P
with n being the total amount of hourly data points, the mean observed
concentration and mean predicted concentration as ¯O = 1
n ∑n
i=1 Oi and ¯P =
1
n ∑n
i=1 Pi
The optimal value of the NMSE lies at 0. It should be noted that the NMSE
is sensitive to extreme outliers as the error is squared. The NMSE was cho-
sen over the RMSE to make comparisons between the different data sets
possible.
Unpaired Peak Prediction Accuracy The unpaired peak prediction accu-
racy (UPPA) is a simple but effective measure of the models accuracy to
predict a days peak concentration. It is defined as the fraction by which the
6
19. 2.3. Verification Statistics
predicted peak concentration differs from the observed peak.
UPPA =
1
days
days
∑
d=1
|max(Od,i) − max(Pd,i)|
max(Od,i)
where the absolute difference of the maximum hourly observed and pre-
dicted concentration |max(Od,i) − max(Pd,i)| is divided by the maximum ob-
served concentration for each day. The UPPA should be as close as possible
to zero.
Fractional Bias A measure of the systematic bias is presented with the
fractional bias (FB). It displays a models tendency to over or under estimate
the concentration.
FB =
∑n
i=1 (Oi − Pi)
0.5 · ∑n
i=1 (Oi + Pi)
A value close to zero indicates the absence of systematic errors, with FB ∈
[−2, 2].
Fraction of predictions within a factor of two The FAC2 is defined as the
fraction of predictions within a factor of two of the observations.
FAC2 =
| Pi
Oi
∈ [1
2 , 2]|
n
The range of this measure is limited to FAC2 ∈ [0, 1] with 1 being optimal.
Balance of correctly predicted episodes versus false and missed episodes
The balance of correct alarms versus false and missed alarms (CFM) is a
statistical measure defined by the author of this study. It displays the mod-
els ability to correctly predict high pollution episodes and penalises falsely
predicted and missed high pollution episodes. A day was considered an
episode if it any of its three hour average concentrations was greater than a
given threshold. The CFM is defined as
CFM = 1 +
C − F − M
C + F + M
where C is the number of correctly predicted episodes, F is the amount of
falsely predicted episodes, M is the number of missed episodes (not pre-
dicted episodes) with CFM ∈ [0, 2]. A CFM value of zero signifies that no
episode was correctly predicted. A CFM value of one means that either
7
20. 2. Methods
the amount of correct predictions was identical to the total amount of false
and missed events combined or that there were no predicted and no actual
events.
The CFM is a generally robust measure. However it is sensitive to the
amount of total actual episodes. If there are no episodes, it becomes only a
measure of false predictions and is limited to maximum 1.
The thresholds were chosen so that all origins had at least 1% of days with
an episode. The thresholds for each pollutant were set identical for all three
stations. NO2 is an extreme case as, in order to reach 1% of days with an
episode at TW and CL, the threshold had to be set so high that the days
with an episode reaches 31% at CB .
For fine suspended particulates the threshold was set to 75 µgm−3, identical
to the daily average value for PM25 suggested by the Hong Kong air quality
objectives in 2014 (HKAQO). The threshold for nitrogen dioxide was set
to 100 ppb, half of the one hour average value for NO2 in the HKAQO.
For Ozone the threshold was chosen to be 50 ppb. The HKAQO suggest a
concentration of 160 ppb as the 8 hour average for O3. The episode criterion
for ozone was set lower in order to observe a sufficient amount of episodes
at CB.
2.4 Procedure
2.4.1 Modeled Bias
As the KFP approximates the true state of a linear variable under noisy
measurements, the variable modelled by the Kalman filter should be linear.
The modelled variable, the model forecast bias, is assumed to be linear. It
is defined to be optimal at zero bias with any value above zero meaning
an over prediction while a value below zero signifies an under prediction.
Three different bias modelling procedures were chosen, the difference (D),
the natural logarithm of the ratio (LR) and the relative difference (RD). These
are explained in the next subsections.
Difference In the case of modelling the direct bias (D), the difference of the
model prediction towards the observation is defined as yt:
yt = Pt − Ot
The observed bias yt is then used to compute a new predicted (true) bias xt
as described in the section 2.1. Finally the predicted bias is combined with
the new model prediction into a Kalman filter corrected prediction PKt+1 :
PKt+1 = Pt+1 − xt
8
21. 2.4. Procedure
Table 2.1: Ratio limits for LR and RD bias. The values are the maximum
ratio of Pi/Oi and Oi/Pi for each origin respectively, restricting less than 1%
of the upper and lower data points.
Ratio PM25 NO2 O3
TW CL CB TW CL CB TW CL CB
max{Pi/Oi} 10 10 10 10 10 10 20 35 45
max{Oi/Pi} 10 10 20 10 10 35 30 25 10
The corrected prediction is controlled to be at least 0.0 as negative concen-
trations do not occur in nature. If PKt+1 is smaller than zero, the absolute
value of PKt+1 is subtracted fromxt and PKt+1 is then set to zero. Like this
the next corrected forecast will be above zero if the new Pt+1 not lower than
the respective Pt. The corrected forecast needs to be controlled less than
1.5% overall for PM25 and NO2. The maximum percentages of controlled
forecasts for ozone concentrations are 7.7% at TW, 6,8% at CL and 15.9% at
CB.
Natural Logarithm of the Ratio Another method of defining the bias yt is
as the natural logarithm of the relationship of model forecasts versus obser-
vations (LR).
yt = log(
Pt
Ot
)
In the case of modelling the relationship we correct:
PKt+1 = Pt+1 · e−xt
Hereby the maximum and the minimum value of yt is controlled in order to
avoid infinite values or unreasonably high values that temporarily bias the
the KF towards an outlier. The limits were set so that less than 1% of the
data points above or below are restricted. The maximum ratios for Pi/Oi
and Oi/Pi were set to 10 and then increased until the requirement was met.
The ratio limits in table 2.1 are used to control yt to stay within the range
[−log(max{Oi/Pi}), log(max{Pi/Oi}].
Relative Difference The last method of defining the bias is the relative
difference (RD). It aims to combine the first two approaches. Here, the bias
yt is defined as
yt =
Pt − Ot
Pt + Ot
and the model forecast is again bias-corrected
PKt+1 = Pt+1 − xt · (Pt+1 + Ot)
9
22. 2. Methods
The optimal RD bias is again zero and falls naturally in the range [−1, 1].
The same ratio limits as explained in the section above and displayed in
table 2.1 are applied so that the range becomes [−max{Oi/Pi}−1
max{Oi/Pi}+1
, max{Pi/Oi}−1
max{Pi/Oi}+1
].
2.4.2 Variations
Two optional variations of the KFP were defined. These were combined into
four total variations: basic KFP, interconnected hours KFP (KF-i), Kalman
filtered variance KFP (KF-v) and both interconnected and Kalman filtered
variance combined (KF-i-v).
Interconnected Hours A simple weighted average is applied to the KFP
input parameters on each iteration to interconnect the separated hours. The
following weighted average was applied:
0.60 · ph + 0.15 · (ph−1 + ph+1) + 0.05 · (ph−2 + ph+2)
where ph is the inputed parameter for hour h. For the first and last hours of
each day the central hour was weighted more instead of using another days
data.
Kalman Filtered Variance In this variation, a Kalman filter is first used to
estimate the true state of the variance σ2
e, t before it is used within the basic
KFP to estimate the state of the forecast bias. The explanation of the detailed
equations for this KF can be found in the appendix B.
Smoothing Optionally, a weighted average of the model data was taken
before it was processed by a KFP. Another optional weighted average was
applied on the corrected forecasts. These data sets are marked by a small
’s’ to symbolise the smoothing. If the weighted average was applied on
unprocessed model data, the hours were weighted with the same coefficients
as the interconnected hours variant.
0.60 · Ph + 0.15 · (Ph−1 + Ph+1) + 0.05 · (Ph−2 + Ph+2)
Additionally a weighted average of the two past days was done:
0.75 · Pd + 0.20 · Pd−1 + 0.05 · Pd−2
The smoothing of the corrected forecasts was done by the application of the
following equations:
0.50 · PKh + 0.20 · (PKh−1 + PKh+1) + 0.05 · (PKh−2 + PKh+2)
0.850 · PKd + 0.125 · PKd−1 + 0.025 · PKd−2
10
23. Chapter 3
Results
The full statistical analysis can not be displayed in this report due to its ex-
tent ( 3 stations ·3 parameters ·18 models · (2 ·12 +3) methods ·5 statistics =
21870 statistical measures). The full analysis of the results of the corrected
model ensemble forecasts can be found in the appendix D.3. It is to be
mentioned that this is only one of the 18 different models. The normalised
average scores of KF corrected forecasts for each model and each method,
averaged over all methods or models respectively, are also included in the
appendix D.2. The most relevant results are the statistical measures for the
best methods found in the next section.
3.1 Performance of Best Methods
The tables 3.1, 3.2, and3.3 show the NMSE, UPPA, FB, FAC2 and CFM of
the selected methods for each pollutant at each station. These are compared
with the performance of the two raw forecasts, their average model ensem-
ble (M-E) and with persistence, which is the time-lagging of observed con-
centrations by one day without any further processing. If the raw model
forecasts were selected as method, the method is labelled ’M’. A model label
of ’Kf-E’ stands for a Kalman filtered ensemble, where the model data is the
average of the KF method applied on both raw forecasts. If the average of all
methods was taken, the model name contains ’Kf-E-mean’. Smoothed data
sets are marked ’s’ and the optional KF variations are ’i’ for interconnected
hours and ’v’ for filtered variance.
11
25. 3.2. Sample Time Series
Table 3.3: Performance of best methods for O3 compared with model perfor-
mance and persistence.
Origin Model Method NMSE UPPA FB FAC2 CFM
TW Kf-E-s-D-i-v-s M 0.412 0.366 0.015 0.739 0.782
CMAQ - 0.972 0.741 0.400 0.552 0.850
CAMX - 0.629 0.448 -0.104 0.565 0.305
M-E 0.619 0.524 0.181 0.619 0.856
Persistence - 0.723 0.487 0.000 0.620 0.734
CL M-E-s D-Kf-i-v 0.367 0.364 0.002 0.763 0.906
CMAQ - 0.830 0.711 0.427 0.630 0.934
CAMX - 0.562 0.507 0.002 0.683 0.476
M-E - 0.564 0.560 0.240 0.700 0.958
Persistence - 0.657 0.450 0.000 0.627 0.834
CB Kf-E-s-LR-v D-Kf-i 0.759 0.587 -0.021 0.626 0.917
CMAQ - 10.988 3.662 1.190 0.133 0.087
CAMX - 3.910 2.359 0.848 0.296 0.203
M-E - 6.564 2.912 1.049 0.193 0.160
Persistence - 1.001 0.688 0.000 0.572 0.615
3.2 Sample Time Series
As TW is the reference station for the Hong Kong SAR, it is used for the sam-
ple time series. Here, the pollutant PM25 is selected because it is the only
data set best approximated by the mean of all KF methods and therefore
should present a balanced exemplary sample. The figures 3.1 and 3.2 are
plots of consecutive predicted and observed concentrations for each hour.
The first figure contrasts the three bias modelling methods and the second
figure compares the four variants for each bias type, here for RD.
13
26. 3. Results
Figure 3.1: Hourly sample plot of PM25 at TW for the three different bias
modelling procedures.
Figure 3.2: Hourly sample plot of PM25 at TW for the four variants of RD.
14
27. 3.2. Sample Time Series
3.2.1 Special Time Series
These figures show how the concentration of a certain hour evolves over a
series of consecutive days. Below each figure the observed bias yt and the
modelled bias xt are plotted for D-, LR- and RD-KF.
Figure 3.3: False model spike corrected by D, LR and RD.
15
28. 3. Results
Figure 3.4: Peaks of observed and predicted concentrations modelled by D,
LR and RD.
16
29. 3.3. Live Display of Actual Implementation
Figure 3.5: Extreme model bias changes modelled by D, LR and RD.
3.3 Live Display of Actual Implementation
Thanks to this study, a basic version of the D-KF was already implemented
at HKUST to automatically correct the AQ model forecasts. The corrected
PM25 forecasts of the next day and the recently passed days for TW, CL and
CB can be observed live at
http://envf.ust.hk/dataview/kfp/current/.
It is planned that the optimised methods will be implemented according to
this study.
17
30.
31. Chapter 4
Discussion and Conclusions
The choice of defining the matrices A, B and C to be identity is a common
practice in most preceding studies. However, in a study by Cheung and
Steenburgh the matrices were all chosen to be identity with the exception of
one station where A was set to 0.75 [1]. It is possible that a value alternative
to identity would have performed better for certain stations in this study,
too. However, a sensitivity analysis of the matrix parameters is out of the
scope of this study.
4.1 Discussion on Best Methods
The overall performance of the best methods for each origin is superior to
the raw model forecasts for NMSE, UPPA, FB, FAC2 and CFM. The only
exception is the CFM for O3 at TW and CL, here the CMAQ model forecast
offers a greater CFM value. In contrast, the prediction of NO2 at all three
stations was significantly improved for CFM.
Remarkable is that even though the raw model forecasts for O3 at CB per-
form several times worse than at the other two stations, the corrected fore-
cast reaches statistical scores comparable to the other stations. This is the
only station where it is recommendable to run a KF twice over the data.
First a LR-KF-v is used to roughly correct the raw model forecasts and then
a D-KF-i is additionally applied to post process the predictions a second
time.
An interaction between NO2 and O3 is expected to be the reason for the poor
performance of the AQM forecast at CB where NO is emitted by the roadside
traffic. The locally increased NO concentration tends to react with ozone,
reducing O3 and increasing NO2 concentrations. The chemical reaction can
be simplified to
NO + O3 → NO2 + O2
19
32. 4. Discussion and Conclusions
and with help of solar radiation the reaction can also occur towards the other
side
NO2 + O2 + hv → NO + O3
Local interactions are hard to model by the AQM, as the spatial resolution
of the model is 1 km.
Further it can be observed that the performance of the KF is not only depen-
dent on the model performance but it is also correlated to the performance
of persistence. NO2 is a good example as the both models perform clearly
worse ad CB while persistence performs slightly better here than at the other
two stations. In the end, the KF corrected forecasts at CB are slightly bet-
ter than at the two other stations. A similar phenomenon can be observed
observed again at CB compared to the two other PM25 stations.
4.2 Discussion on Time Series
The time series are too short to draw reliable conclusions about the over-
all performance about the respective KF. The selected statistics measure the
general performance more reliably as a large amount of data was evaluated.
Nevertheless, an analysis of the time series is important to understand the
strengths and weaknesses of the KF. From figure 3.1 several insights can be
won. First, it can be seen that the three bias modelling techniques difer in
their peak correction performance. The LR-KF correctly decreases the false
model peak at hour 20 most compared to the D-KF which follows the model
predictions more closely. The RD-KF lies somewhere in between. At hour
45 the bias changes from a strong over prediction to a strong under pre-
diction. Now the KFP all perform worse than the raw forecast. The D-KF
even reaches its lower limit for several consecutive hours because its bias
correction is not relative to the magnitude of the forecast concentration. All
the methods converge in the following days again towards the observations.
Therefore, the second and third insights are that a sudden change of bias
leads to temporarily worse performance of the KF, but it then adapts to the
new situation and changes its bias correction in response to the new ob-
served bias. Whereas the rate of adaptation depends linearly on the degree
of discrepancy between the observed and modelled bias.
Figure 3.2 suggests that the differences between the optional variations are
rather small. Generally it can be seen that the variant filtering the variance
is more volatile than the interconnected variants.
Special Time Series As a separate KF is run for each hour, the special
time series plotting one hour over few days are most helpful to understand
how the KF works and to compare the different bias modelling techniques.
The falsely over predicted peak concentration seen in figure 3.3 is correctly
20
33. 4.3. Conclusions
downwards corrected by all KF methods, while the magnitude of correction
differs. D-KF follows the model closer while LR-KF predicts the lowest peak.
The estimated D bias state adapts slowly to the fast occurring peak at day
104 while the LR and RD bias are both estimated almost perfectly at day 104.
Nevertheless, the RD-KF corrected forecast differs from the LR correction
because its correction magnitude depends linearly on the last observation
whereas the LR correction only depends on the new model prediction and
the estimated LR bias (see bias definitions in the methods section 2.4.1).
Altering the weight of the past observation on the correction magnitude
could improve the performance. For example as: yt = Pt−Ot
Pt+0.5·Ot
and PKt =
Pt+1 + xt · (Pt+1 + 0.5 · Ot)
All methods fail to predict the first observed peak in figure 3.4 due to miss-
ing indication of a rise by the AQ model. The Kalman filter is not able to
create additional information, but can only un-bias the forecast. As a cause
of the first not predicted peak, the second peak was over predicted, partially
also because the observed concentration did not rise proportionally to the
predicted concentration. Finally, the last peak is best predicted by the LR-KF,
whereas it also most over predicted the second peak.
Figure 3.5 shows an extremely volatile observed bias. None of the Kalman
filtering methods are able to successfully adapt to the fast changing bias.
Therefore all three methods estimate the true bias to lie around zero. The
extreme bias ratio at day 166 is one of the cases where the ratio is limited
to the maximum ratio, defined in table 2.1, in order to prevent the KF from
being overly influenced by this outlier in the next time steps.
4.3 Conclusions
Each origin is best corrected by a different KF method. The performance of a
specific KF implementation depends on the characteristics of the underlying
model bias particular to each station. An exception is the prediction of NO2
where the RD bias methods are most skilful at all stations. Generally, the
LR-KF is best at correcting extreme differences and peaks, whereas is also
tends to over predict peaks. The D-KF and RD-KF perform better in cases
with a moderate AQ model bias.
For most stations a KF corrected ensemble of both raw model forecasts per-
formed best. PM25 prediction at TW is even optimally corrected by the
average of all introduced methods and variants. Post processing the fore-
casts twice is only recommendable for data sets with an extreme model bias
such as for O3 at CB. The type of correction done to the raw forecasts can be
described in a simplified manner as a time-varying additive correction for
the D-KF and as a time-varying multiplicative correction for the LR-KF. The
RD-KF is a combination of both.
21
34. 4. Discussion and Conclusions
The comparison with persistence shows that the raw AQ model forecasts
perform worse than a simple time-lagging of observations. The KF corrected
forecasts in contrary are more skilful. Therefore the AQ model provides
relevant information that can be used to generate accurate forecasts, but a
skilful MOS such as the KF is necessary to remove the bias.
The Kalman filters ability to adapt to a changing bias without any knowl-
edge of the underlying model makes it an attractive post processing method
to use with AQ data which does not always follow expected physical pat-
terns due to complex interaction. Additionally its recursive implementation
makes day to day operation simple and does not require a lot of computa-
tion power.
4.4 Recommendations
In this study a fixed ratio of variances r = 0.1 was applied. A sensitivity
analysis of the ratio r and the definition of an optimal value for each sta-
tion and pollutant is recommended. Values different from identity for the
matrices A, B and C could be tested. Further sensitivity analysis’ could be
done on the coefficients used in the different weighted averages and on the
amount of past days to compute the variance.
As the combination of the raw forecasts generally showed an improved per-
formance, an extended ensemble forecast consisting of multiple AQ model
runs with different scenarios combined into one ensemble should be consid-
ered.
As the a logarithmic ratio bias modelling technique also showed good per-
formance for certain data sets, the extended Kalman filter (EKF) for non
linear systems could promise an increased performance.
Different approaches of combining a KF with a traditional MOS such as
a multiple regression (MR) can be explored. The possibility of combining
both in one overlapping step, such as a Kalman filtering of the regression
parameters is to be explored. A spatial correlation analysis could overcome
part of the systematic bias due to the limited spatial resolution. It is to be
expected that a MOS combination with the KF applied last performs best as
it is highly adaptive and can correct even temporary systematic bias.
Finally, an intelligent AQMOS system could utilise the strengths of the differ-
ent KF methods presented in this study. This could be done by running a set
of KF methods for each station which are constantly analysed for their per-
formance. Then the forecast corrected by the recently most skilful method
would be selected. Or different situations types such as high and low pollu-
tion, volatile and steady bias could be defined and then the KF method best
performing in the current situation type selected for the forecast correction.
22
35. Appendix A
A short introduction to the Kalman
filter
The Kalman Filter is a state of the art estimator of the true state of a variable
affected by noisy measurements. It is commonly used to process signals and
can be found in a large variety of applications. A popular example would
be a GPS device estimating its true position from signals affected by random
errors.
The true (unknown) state is modelled by the KF and updated each time a
new signal is received. The magnitude of each correction depends on how
large the recent random error of the signal is.
First, the (true) state and the expected signal are predicted by
ˆxt = Axt−1 + Bµt + ηx
ˆzt = C ˆxt + ηz
where the predicted state ˆxt is dependent on the last state xt−1 and a system
variable µt and affected by a normal distributed random error ηx, with zero
mean and variance Ex. The sensor prediction differs from the state by a
normal distributed random error ηz, with zero mean and variance Ez. A, B
and C are matrices describing the linear relationship between the variables.
The state covariance matrix ˆRt, describing the expected square errors of the
modelled system parameters, is estimated to be
ˆRt = ARt−1AT
+ Ex
Next, the Kalman gain Kt can be computed as
Kt =
ˆRtCT
C ˆRtCT + Ez
23
36. A. A short introduction to the Kalman filter
Now the state xt and its covariance matrix Rt are updated
xt = ˆxt + Kt(zt − ˆzt)
Rt = (I − KtC) ˆRt
Where zt − ˆzt = zt − C ˆxt is the difference of the sensor prediction to the
actually received signal. I is the identity matrix. The Kalman gain Kt ∈ [0, 1]
is a measure for the trust the KF puts into the new signal and how much
the state is corrected towards the signal. The gain depends on the changing
variances and is different for each time step.
This recursive algorithm is then repeated for each time step. In the case of
estimating the true state of the AQ forecast bias, the measurement for the
future day is not yet available. Therefore the Kalman Filter has to be used
in predictor mode, sometimes also referred to Kalman Filter Predictor (KFP)
which will be discussed in the methods chapter in section 2.1.
24
37. Appendix B
Variance Computation and Initial
Values
Preceding studies stated that the KFP performance is independent of the
actual value of the variances but is sensitive to the ratio of the variances. A
fixed variance was used in those studies [2, 1]. However, in this study it was
chosen to use a time-varying variance as some early trials showed greater
performance for this approach. The ratio of the variances was also fixed to
r =
σ2
q
σ2
e
= 0.1
Strongly simplified, the ratio r can be thought of as a minimal value of the
Kalman gain. The variance σ2
q will be simply computed as σ2
q = r · σ2
e . the
ratio of the variances used in the KFP modelling the variance of et can be
expressed as
σ2
σ2
q
σ2
σ2
e
= r2
The variance σ2
σ2
q
is therefore be expressed as r2 · σ2
σ2
e
The basic KF’s time-varying variance σ2
e is computed from the variance of the
recent errors et = yt − xt. As the Kalman filter remembers the systematics
of the past 4 to 5 days, it was chosen to calculate the variance from the
past 5 data points. During the first five spin-up days the variance is simply
computed from the yet available data points.
The variance σ2
σ2
e
used in the Kalman filter estimating the variance σ2
e, t is
computed by calculating the variance of the Kalman filtered variance σ2
e, t of
the past five days.
It should be noted that the variance calculated for the basic KFP and the
Kalman filtered variance are not the same. The variance of the first con-
25
38. B. Variance Computation and Initial Values
tains both systematic and random components whereas the second should
exclude the random error components.
Kalman Filtered Variance Here the equations applied in the KF variant
which additionally models the variance are described. The Kalman gain
used estimate the true state of the variance σ2
e, t is computed as
Kt =
Rσ2
e , t−1 + σ2
σ2
q
Rσ2
e , t−1 + σ2
σ2
e
+ σ2
σ2
q
then the state of the variance σ2
e, t and its expected mean square error Rσ2
e , t
are updated
σ2
e, t = σ2
e, t−1 + Kt(ˆσ2
e, t − σ2
e, t−1)
Rσ2
e , t = (1 − Kt)(Rσ2
e , t−1 + σ2
σ2
q
)
where the estimated signal ˆσ2
e, t of the variance σ2
e, t is approximated using
the relationship of the variances. From the time series
zt = yt − yt−1 = xt + et − xt−1 − et−1 = qt−1 + et − et−1
follows that σ2
z = σ2
q + 2σ2
e = (2 + r)σ2
e and σ2
e = σ2
z
2+r Therefore we can
approximate
ˆσ2
e, t =
(yt − yt−1)2
2 + r
Initial Values The initial values for the estimated state xt and σ2
e, t of the
basic KFP and the KFP modelling the variance were simply set to zero. The
initial values of the variances σ2
e , σ2
σ2
e
and the expected mean square errors Rt,
Rσ2
e , 2 however needed to be set to a positive value for the KFP to converge
to an optimal solution. The initial values of σ2
e and Rt were chosen to be the
square of the first observed bias y2
1. Analogously the initial values of σ2
σ2
e
and
Rσ2
e , 2 were set to be the square of the initial variance σ4
e, init = y4
1. Hereby we
assume y1 not to be zero.
26
39. Appendix C
Literature Review
AQMOS: Air Quality Model Output Statistics from CMAQ model forecasts
In this publication of 2010, Miller et al. [7], aims to demonstrate how the
bias of air quality model output can be adjusted to generate more accurate
predictions. The demonstrated benefit of the AQMOS is the reduction of the
systematic model error (bias) for site specific predictions using observation
data.
A daily regression is computed between recent historic Air Quality Model
(AQM) predictions and observations. For each city, pollutant and model spe-
cific regression equations are employed. Further a division between normal
and high pollution days is made. A clear systematic bias can be seen on
a scatter plot of observed versus model data. Whereas high pollution days
data shows a different bias from low pollution days data. Therefore two re-
gression equations are created, one for normal days using all data and one
for high pollution days using only data above the 85th percentile of model
output.
As an alternative to regression analysis, two simple bias-adjustment strate-
gies, mean-bias and multiplicative ratio are named. As an advanced strategy,
the Kalman Filter Scheme by Delle Monache et al. 2006 [2] is pointed out.
This scheme is reviewed later.
Miller et al. measure the performance by comparing the average difference
of the raw model output and the bias adjusted data towards the observed
measurements. A particular focus is set on days that are critical air quality
days. The prediction of these improved in most cities through AQMOS.
A review on the forecasting of wind speed and generated power Wind
speeds posses in most sites a high temporal fluctuation. As the generated
power of wind turbines is directly proportional to the prevailing wind speed,
the prediction of wind speeds is crucial. If the generated power can be
27
40. C. Literature Review
predicted more accurately, the electricity generation can be optimized for
the system load without generating excess electricity.
Lei et al. give in this paper of 2009 [5] a brief survey over the most used and
some new methods to forecast wind speed and power. For long term pre-
diction physical methods perform better, whereas for short-term prediction
statistical methods are preferable. Often both methods are utilized simulta-
neously. Further, the spatial correlation model is also viewed to be relevant.
Artificial intelligence and other new methods have gained growing attention
for geophysical prediction.
As model input the physical models utilize physical and meteorological in-
formation. Statistical and spatial correlation models need NWP output and
historic data as input, whereas the later also needs the data from neigh-
bouring sites. A Spatial Correlation Predictor has been developed by Alexi-
adis, 1998 [6] to benefit from the spatial relationship of different sites wind
speeds.
To improve the predictions of physical models, accurate Digital Elevation
Models (DEMs) can be applied to the Numerical Weather Prediction (NWP)
and Model Output Statistics (MOS) are commonly used to post-process the
model output data.
Conventional statistical models include Autoregressive (AR), Moving Aver-
age (MA), and Autoregressive (integrated) Moving Average (AR(i)MA). The
equation C.1 shows the ARMA model. Note that for m = 0 the model
becomes the AR model.
xt =
n
∑
i=1
ϕixt−i + αt −
m
∑
j=1
θjαt−j (C.1)
ϕ AR parameter
θ MA parameter
α Normal distributed white noise
Other mentioned procedures cover ANN (Artificial Neural Network), MLP
(Multiple Layer Neural Perception Neural Network), LF-DFNN (Locally
Feedback Dynamic Fuzzy Neural Network) and SVM (Support Vector Ma-
chine).
Statistical post processing of model output from the air quality model
LOTOS-EUROS Annemiek Pinjappel uses R in her work of 2011 [8] to im-
plement MOS. The performance is measured by R2, RMSE as well as missing,
false and well modelled alarms. She found that for PM10 the LOTOS-EUROS
28
41. model output improved substantially. However, the division of the data into
summer and winter time did not lead to a significant improvement.
Combining deterministic and statistical approaches for PM10 forecasting
in Europe Knovalov et al. found in their publication of 2009 [4] that the
improvement is better the more variables are in the regression. The most
important meteorological parameter in summer is the temperature and in
winter the boundary layer height. The cross-validation method is used for
the performance measurement. The result of the study is that MOS lead to
a significant improvement not only for a given site but also for territories of
similar environment far away from the considered site.
Ozone ensemble forecasts: 2. A Kalman filter predictor bias correction In
2006, Luca Delle Monache et al. test the ability of the Kalman Filter in pre-
dictor mode to reduce systematic errors. Ozone forecasts are postprocessed,
using the bias of past forecasts to bias-adjust the future predictions. The
study finds the adjusted forecasts to become better considering gross error,
RMSE, correlation and to predict peaks far more accurate.
Figure C.1: Flow chart describing the Kalman filter predictor process,
adapted from Delle Monache et al.[2]
In the previous part of the study, a Ozone Ensemble Forecast System (OEFS)
was introduced. 12 forecasts were computed with different meteorological
inputs and emission scenarios. The mean of the ensemble performed better
than the individual forecasts. All forecasts exhibit a systematic error - a
typical problem in CTMs (Chemical Transport Models) [9]. In this part of
the study, the KF is applied on each forecast to minimize its systematic bias.
The KF is suited for this task. Due to its recursive and adaptive methodology
it can also include the temporal variation of the forecast error.
Figure C.1 shows how the KF is used to predict and correct. The bias is
29
42. C. Literature Review
constantly being re-estimated. The bias ( ˆxt t−∆t), estimated at the previous
timestep, is rectified by a fraction (β) of its discrepancy towards the observed
forecast error (yt) and then used to predict the next timeseteps bias ( ˆxt+∆t t).
30
43. Appendix D
Statistical Meaures
D.1 Normalisation of Statistics and Selection of Best
Methods
In order to be able to compare the performance of the different methods, the
statistical measures were normalised and combined into a single score for
each method and model combination at each origin. The score was designed
to represent the best method at a value of 1 and the worst method would
receive a score of 0. The normalised value of NMSE, UPPA and FB was
computed as
˙Si = 1 −
Si − min(Si)
max(Si) − min(Si)
and for FAC2 and CFM as
˙Si =
Si − min(Si)
max(Si) − min(Si)
where ˙Si is the normalised statistical measure Si. The Si of the FB was taken
as its absolute value |Si| because any positive or negative deviation from
zero represents a bias.
Then the average of the five normalised measures ˙Si was calculated for each
method and model combination and compared within the same dataset of
an origin. The method and model combination with the highest score was
then selected to be optimal for the respective origin. This procedure was
repeated to analyse only data points over or below a certain percentile to
gain further insight over the data. Due to the limited extent of the study,
these additional results will not be discussed.
31
55. Bibliography
[1] William Cheung and James Steenburgh. Strengths and weaknesses of
mos, running-mean bias removal, and kalman filter techniques for im-
proving model forecasts over the western united states. Weather and
forecasting, 22(6):1304–1318, 1986.
[2] Luca Delle Monache, Xingxiu Deng, Thomas Nipen, Roland B. Stull, and
Yongmei Zhou. Ozone ensemble forecasts: 2. a kalman filter predictor
bias correction. Journal of Geophysical Research Atmospheres, 111(D5):2156–
2202, March 2006.
[3] Ian T. Jolliffe and David B. Stephenson. Forecast verification:a practi-
tioner’s guide in atmospheric science. December 2011.
[4] I.B. Konovalov, M. Beekmann, F. Meleux, A. Dutot, and G. Foret. Com-
bining deterministic and statistical approaches for pm10 forecasting in
europe. Atmospheric Environment, 43(40):6425 – 6434, 2009.
[5] Ma Lei, Luan Shiyan, Jiang Chuanwen, Liu Hongling, and Zhang Yan. A
review on the forecasting of wind speed and generated power. Renewable
and Sustainable Energy Reviews, 13(4):915–920, 2009.
[6] Alexiadis MC, Dokopoulos PS, Sahsamanoglou HS, and Manousaridis
IM. Short-term forecasting of wind speed and related electrical power.
Solar Energy, 63(1):61–80, 1998.
[7] Dianne S. Miller, Clinton P. MacDonald, Timothy S. Dye, Kenneth J.
Craig, and Daniel M. Alrick. Aqmos: Air quality model output statis-
tics from cmaq model forecasts. 9th Annual CMAS Conference, October
2010.
[8] Annemiek Pijnappel. Statistical post processing of model output from
the air quality model lotos-euros. Stegeverslag, 2011.
43
56. Bibliography
[9] Armistead Russell and Robin Dennis. Narsto critical review of photo-
chemical models and modeling. Atmospheric Environment, 34(12):2283–
2324, 2000.
44