Marco Cattaneo "Event data processing in LHCb"Yandex
Семинар «Использование современных информационных технологий для решения современных задач физики частиц» в московском офисе Яндекса, 3 июля 2012
Marco Cattaneo, CERN
Marco Cattaneo "Event data processing in LHCb"Yandex
Семинар «Использование современных информационных технологий для решения современных задач физики частиц» в московском офисе Яндекса, 3 июля 2012
Marco Cattaneo, CERN
Niko Neufeld "A 32 Tbit/s Data Acquisition System"Yandex
Семинар «Использование современных информационных технологий для решения современных задач физики частиц» в московском офисе Яндекса, 3 июля 2012
Niko Neufeld, CERN
Overview of Photonics Research at Calit2: Scaling from Nanometers to the EarthLarry Smarr
10.03.26
UCSD-NICT Joint Symposium on Innovative Lightwave,
Millimeter-Wave and THz Technologies
for Future Sustainable Network
Title: Overview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
La Jolla, CA
DSD-INT 2020 Real Time Hydrologic, Hydraulic and Water Quality Forecasting in...Deltares
Presentation by Tony McAlister, WaterTech, at the Delft3D User Days - Australian Time zone: Inland to Estuary, during Delft Software Days - Edition 2020. Tuesday, 10 November 2020.
Opensample: A Low-latency, Sampling-based Measurement Platform for Software D...Junho Suh
In this paper we propose, implement and evaluate OpenSample: a low-latency, sampling-based network measure- ment platform targeted at building faster control loops for software-defined networks. OpenSample leverages sFlow packet sampling to provide near–real-time measurements of both net- work load and individual flows. While OpenSample is useful in any context, it is particularly useful in an SDN environment where a network controller can quickly take action based on the data it provides. Using sampling for network monitoring allows OpenSample to have a 100 millisecond control loop rather than the 1–5 second control loop of prior polling-based approaches. We implement OpenSample in the Floodlight OpenFlow controller and evaluate it both in simulation and on a testbed comprised of commodity switches. When used to inform traffic engineering, OpenSample provides up to a 150% throughput improvement over both static equal-cost multi-path routing and a polling-based solution with a one second control loop.
The Remarkable Benefits and Grave Dangers of using Artificial Intelligence in...Steve Cuddy
Overview
What is Artificial Intelligence (AI)
Petrophysical Case Studies showing successful applications
- Evolution of shaly water saturation equations
- Nuclear Magnetic Resonance T1 & T2 spectra analysis
- Prediction of shear velocities
- Litho-facies and permeability prediction
- The log quality control and repair of electrical logs
Narrow vs. General vs. True AI
The grave dangers of using AI
- More than AI making poor petrophysical predictions!
- I describe an end of civilisation scenario
A Deep Learning use case for water end use detection by Roberto Díaz and José...Big Data Spain
Deep Learning (DL) is a major breakthrough in artificial intelligence with a high potential for predictive applications.
https://www.bigdataspain.org/2017/talk/a-deep-learning-use-case-for-water-end-use-detection
Big Data Spain 2017
November 16th - 17th Kinépolis Madrid
This poster was presented at the 2015 Water Reuse Symposium in Seattle and represents ADMi's analysis of data to determine which water quality parameters have the greatest impact on RO membrane fouling.
Niko Neufeld "A 32 Tbit/s Data Acquisition System"Yandex
Семинар «Использование современных информационных технологий для решения современных задач физики частиц» в московском офисе Яндекса, 3 июля 2012
Niko Neufeld, CERN
Overview of Photonics Research at Calit2: Scaling from Nanometers to the EarthLarry Smarr
10.03.26
UCSD-NICT Joint Symposium on Innovative Lightwave,
Millimeter-Wave and THz Technologies
for Future Sustainable Network
Title: Overview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
La Jolla, CA
DSD-INT 2020 Real Time Hydrologic, Hydraulic and Water Quality Forecasting in...Deltares
Presentation by Tony McAlister, WaterTech, at the Delft3D User Days - Australian Time zone: Inland to Estuary, during Delft Software Days - Edition 2020. Tuesday, 10 November 2020.
Opensample: A Low-latency, Sampling-based Measurement Platform for Software D...Junho Suh
In this paper we propose, implement and evaluate OpenSample: a low-latency, sampling-based network measure- ment platform targeted at building faster control loops for software-defined networks. OpenSample leverages sFlow packet sampling to provide near–real-time measurements of both net- work load and individual flows. While OpenSample is useful in any context, it is particularly useful in an SDN environment where a network controller can quickly take action based on the data it provides. Using sampling for network monitoring allows OpenSample to have a 100 millisecond control loop rather than the 1–5 second control loop of prior polling-based approaches. We implement OpenSample in the Floodlight OpenFlow controller and evaluate it both in simulation and on a testbed comprised of commodity switches. When used to inform traffic engineering, OpenSample provides up to a 150% throughput improvement over both static equal-cost multi-path routing and a polling-based solution with a one second control loop.
The Remarkable Benefits and Grave Dangers of using Artificial Intelligence in...Steve Cuddy
Overview
What is Artificial Intelligence (AI)
Petrophysical Case Studies showing successful applications
- Evolution of shaly water saturation equations
- Nuclear Magnetic Resonance T1 & T2 spectra analysis
- Prediction of shear velocities
- Litho-facies and permeability prediction
- The log quality control and repair of electrical logs
Narrow vs. General vs. True AI
The grave dangers of using AI
- More than AI making poor petrophysical predictions!
- I describe an end of civilisation scenario
A Deep Learning use case for water end use detection by Roberto Díaz and José...Big Data Spain
Deep Learning (DL) is a major breakthrough in artificial intelligence with a high potential for predictive applications.
https://www.bigdataspain.org/2017/talk/a-deep-learning-use-case-for-water-end-use-detection
Big Data Spain 2017
November 16th - 17th Kinépolis Madrid
Similar to Wqtc2011 causes offalsealarms-20111115-final (20)
This poster was presented at the 2015 Water Reuse Symposium in Seattle and represents ADMi's analysis of data to determine which water quality parameters have the greatest impact on RO membrane fouling.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
A tale of scale & speed: How the US Navy is enabling software delivery from l...
Wqtc2011 causes offalsealarms-20111115-final
1. Causes of False Alarms in
Inferential Event Detection
Systems for Distribution System
Water Quality Monitoring
Ed Roehl, John Cook, Ruby Daamen, and Uwe Mundry
Advanced Data Mining International, LLC
Greenville, South Carolina
2. Early Results from
WRF PROJECT 4182
Interpreting real-time online
monitoring data for water
quality event detection
3. Acknowledgements
ADMi gratefully acknowledges the Water
Research Foundation as the joint owners of
certain technical information upon which this
presentation is based. ADMi thanks the
Foundation for their financial, technical, and
administrative assistance in funding the
project through which much of this
information was discovered.
4. Thanks to Our Utility Partners
• City of Columbus, Ohio, Division of Power
and Water
• Greenville Water System
• Newport News Waterworks
• Oklahoma City Water Department
• Startex Jackson Wellford Duncan Water
District (SJWD)
6. Inferential Event Detection System (IEDS)
• Focus on distribution system security
• Real-time monitoring of “conventional” WQ
parameters - CL2, PH, COND/SC, TURB,
TOC, TEMP
• Infers an “event” by detecting anomalous
patterns of WQ behavior
– Does not measure concentrations of specific
compounds like liquid chromatography
• Systems have been available for a number
years
7. Colorado State pilot loop from Project 3086
Hach “panel” SC, pH, Cl2, turbidity TOC
data acquisition analyzer
ventilation
toxin
flow injection
pump pump
flow direction flow loop injection point
8. CSU pilot loop results
chlorine residual chlorine residual response
response to Aldicarb to Na Cyanide
9. conductivity CSU pilot loop results, cont.
pH
conductivity response to pH response to Na
Na Arsenate Cyanide
10. HOW? - event detection
event
SCADA historical database of
1 “normal” behaviors”
5
feature vector 6
CL2 TURB distances to
dCL2 pH SC
neighbors
2 dt TOC
CL2
4
vector
track
3D projection of n-dimensional
feature space 3
12. Measurement errors, tank cycling, etc.
1 week
data
gap
10-minute time steps
• Fast, full scale change
13. WRF Project 4182
• Reports of unacceptable numbers of false
positives unless sensitivity reduced
– defeats purpose
• Thesis - a more effective IEDS can be
developed by incorporating the effects of
operational parameters on water quality
variability
– reduce false positives
– local ops params – Q, Ps, LVL
• Utility partners provided multi-year data from
40+ monitoring sites
14. Definitions
• A contaminated slug flowing past a
sensor array might only be detectable for
a few minutes or less.
– Here, target detection window 20 minutes
• event is manifest and detectable
• “Normal” data – all the data here
represents normal operations and
normal data collection issues.
16. Question 1
• Q1: If event detection relies on pattern matching vectors,
how similar are “normal” new vectors to normal old
vectors?
– “old” = historical database
– “new” = vectors streaming from process
• Expected A: If both are normal, they should be pretty
similar.
CARTOON
historical normal new normal data
data
17. Experiment 1 – determine if old and new vectors
cohabitate the same sub-spaces of feature space
CL2 (mg/l)
1. Divide 4 years
historical new of 10-min data
into ~70% old
and ~30% new
2. Define vector
scalar CL2
features for
each WQ
parameter
coarse a. Scalars -
divided into 5
segmentation 20% sub-ranges
18. cont 1: Experiment 1 - cohabitating hist. and new
2. cont. - Create
coarse
features
segmentation
b. D1 = 1-time-
step difference;
sub-divide into
6 sub-ranges
c. D2 = D1 time-
delayed 1 time
step
D1 & D2 CL2
Process Dynamics – scalar+D1+D2 describe
parameter’s current position+velocity+acceleration.
19. cont 2: Experiment 1 – cohabitating old and new
3. Count cohabitating old and new in sub-spaces
(hypercuboids) formed by 5 scalar, 6 D1, and 6 D2 sub-
ranges
Combinatorial Explosion – even with coarse segmentation
- 3 scalars = 5 x 5 x 5 = 125 cuboids
- + D1 = 125 x 6 x 6 x 6 = 27k hypercuboids
- + D2 = 27k x 6 x 6 x 6 = 5.8 million
tank site
20. cont 3: Experiment 1 – cohabitating hist. and new
booster pump station away from tanks
next experiment
Back to Question 1
• Q1: How similar are “normal old” and “normal new”
vectors?
• Expected A: If both are “normal”, they should be
pretty similar.
• Real A: Not very - numerous false alarms may be
unavoidable without desensitizing IEDS
21. Question 2
• Q2: What would happen if we periodically transfer
“new” vectors to the historical database?
– Experiment 1 – static old & new
• Expected A: False alarms should decrease.
22. Experiment 2 – simulate updating hist. database
new vectors 47,624
• Simulations used the site away from tanks
– features = CL2, SC, COND, TURB scalars+D1s+D2s
• dnn = distance of new vector to “nearest neighbor” old vector
– In IEDS dnn > specified limit triggers alarm
• Findings
1. Transfer cases are high percentage of no-transfer case
2. Little difference between transfer cases
3. Indicates that successive “normal” vectors can be far apart
24. Back to Question 2
• Q2: What would happen if we periodically
transfer “new” vectors to the historical
database?
• Expected A: False alarms should decrease.
• Real A: False alarms might not fall to
acceptable levels.
25. Question 3
• Q3: Why are successive vectors so
far apart?
• A: To come.
26. Experiment 3 – correlation matrices
• Cross-correlation matrix – correlates changes among multiple
ops & WQ parameters
– change = Dx = current value – value x time steps ago
• Utility B stand alone
site
• 86-sec time step 1 time-step
• Mix of WQ and
operational
parameters
3 time-step
(4.3 min)
change
7 time-step
(10 min)
change
27. Experiment 4 – autocorrelation of Dx
• Autocorrelation function correlates a signal to itself to
determine how deterministic / random it is.
– determinism = current behavior depends somewhat on past
– randomness = current behavior unrelated to past
1 time-step
14 time-step
(86 sec)
(20 min)
change
change
28. Back to Question 3
• Q3: Why are successive vectors so far
apart?
• A: WQ change on time scales 20 minutes
can be “apparently random”.
– Exp. 3 (x-matrices) - WQ & ops parameter
changes are poorly correlated
– Exp. 4 (autocorr.) - individual WQ parameter
change is non-deterministic
– Same findings at multiple sites & utilities
• non-determinism = randomness = noise
29. Causes of WQ variability
Stand
Alone
Site
• Unmeasured disturbances
– pressure & flow transients
• Measurement errors
31. Alternative to stand-alone site
monitoring
Event
Detector
• Upstream / downstream sites
• Upstream site provides
– boundary conditions for downstream WQ
– more operational parameters
32. Multi-Site Concept
• Event detection performed on filtered signals
– model-based filtering of downstream WQ signals
– modeling = accounting of causes of variability
– filtered signals less variable
• Modeling technique
– multivariate, nonlinear curve fitting by (multi-layer
perceptron) artificial neural networks (ANN)
• “machine learning” from AI
– inputs - upstream and “local” WQ and ops
• spectrally decomposed into components
• autoregressive “local” WQ inputs time delayed to be outside
detection window (e.g., 20 minutes)
• co-linear inputs decorrelated
– ANN “learns” best predictor components
33. upstream
flow
downstream 2-Site Example
COND (mS/cm) TEMP (deg. F)
COND
TEMP
test data
PH
CL2
CL2 (mg/l)
PH
1-hour time steps (220 days, August to March)
• Raw WQ variability is similar but not identical
– differences caused by unmeasured disturbances
• 1-hour time step too big for 20-minute detection window
– exploratory research on multi-site
34. Results – detail
Downstream COND D1 (mS/cm)
D1 = 1 time-step difference
of test data
measured data
Downstream CL2 D1 (mg/l) upstream only
upstream+auto
Downstream PH D1
1-hour time steps
35. More Complicated 4-Site Example
Q? = unmeasured
LVL,
disturbances COND, TANK
CL2
Q, PSUC, PDIS,
COND, CL2, TEMP
A Q, PSUC, PDIS,
COND, CL2, TEMP
BPS BPS
B A
LVL, Q? = unmeasured
TANK COND, disturbances
CL2
B
• BPS B is “target” site
• Utility operates multiple WTPs with different
sources
• 1 year of data (1-min reduced to 4-min)
– first 10 month = training
– last 2 months = test
36. BPS B COND Process Model – training data
Training Data
measured predicted residuals N: 76,148
R2: 0.847
RMSE: 72 mS/cm
Residual Error (mS/cm)
BPS B COND (mS/cm)
4-minute training data observations
• Looks Good!
37. BPS B COND Process Model – test data
Test Data hump
N: 17,296
R2: 0.893
RMSE: 69 mS/cm
Residual Error (mS/cm)
BPS B COND (mS/cm)
measured predicted residuals
4-minute test data observations
• Hump may be from different WTP/source
• Looks Good!
38. BPS B COND Process Model – test data
Detail
BPS B COND (mS/cm)
measured predicted
4-minute test data observations
• Looks Bad!
• Process model misses some periods - maybe
from unmonitored flows through junctions
39. BPS B COND Dx autocorrelations
COND
• Dx = D
R R2 over x
number of
minutes
CL2 • R2s are
low
R R2
40. BPS B CL2 Process Model – test data
trough
Residual Error (mg/l)
BPS B CL2 (mg/l)
Test Data
measured predicted residuals N: 11,715
R2: 0.912
RMSE: 0.085 mg/l
4-minute test data observations
• trough may be from different WTP/source
• Looks Good!
41. BPS B CL2 Process Model – test data
Detail
BPS B CL2 (mg/l)
measured predicted
4-minute test data observations
• Looks Bad!
• Process model missing some periods - maybe
from unmonitored flows through junctions
43. IEDS - Conclusions
• Practical problems
– data reliability
– no guarantees that contamination event would “look”
different than “normal” because
• “normal” is so highly variable
• WQ sensors being used might not provide the “information”
necessary to discriminate
– Where to put / how many?
• Stand-Alone Sites
– face widely ranging random variability from unknown
disturbances, a.k.a. normal operations
– high alarm limits needed to reduce false positives -
defeats purpose
44. cont - Conclusions
• Multi-Site approach
– Can account for/explain 80-90% of downstream
WQ variability
– unproven on 20 min detection window
– diminished when too much complexity
– field testing to be done at GWS and SJWD
45. cont - Conclusions
• Other reasons to monitor distribution system WQ
– control processes at WTP to improve WQ at points of delivery
– detect common problems - low total chlorine, nitrification, line
integrity, DBPs, biofilm sloughing, incipient complaint detection
CL2 (mg/l)
BPS A TANK A
BPS B TANK B
1-minute time steps 1/1/05 – 11/16/09
46. Compare WTP with DS turbidity
• Little variability in WTP turbidity, < 0.1 NTU !
47. Correlate DS turbidity with WTP WQ
• ANN process model
• Inputs = finished water
alkalinity, hardness, color, and
source blend ratio
• R2 = 0.71