International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
DETERMINATION OF NET FLOWS INTO ALMATTI RESERVOIR FROM CWC GAUGE DATA AND RES...IAEME Publication
This paper presents the determination of net flows into Almatti reservoir from CWC data and reservoir data. From the study it can be concluded that the average flow in to Almatti will be 574.86 TMC, the maximum inflow will be 1196.8 TMC and the minimum flow will be 166.99 TMC. The flows in annual in deficit years may reduce by about 50 TMC but there is no variation in the good years in the good years as the storage effects will take care of this aspect during good years. It can be concluded that there will be reduction of flows in the June and July flows in the ultimate scenario except in very good years.
Floodplain Modelling Materials and MethodologyIDES Editor
A floodplain is the normally dry land area adjoining
river or stream that is inundated during flood events. The
most common reason for flooding could be overtopping of river
or stream due to heavy downfall. The floodplain carries flow
in excess of the river or stream capacity. Flood frequency and
flood water-surface elevations are the crucial components for
the evaluation of flood hazard. This paper presents the
methodology that incorporates advanced technologies for
hydrologic and hydraulic analyses that are needed to be carried
out to predict the flood water-surface elevations for any
ungaged watershed.
Determination of safe grade elevation by using hec ras case study mutha rivereSAT Journals
Abstract
Flood is a naturally occurring disastrous event causing damages, losses and destruction to property, life and environment.
Hundred millions of money are spent every year in flood control and flood forecasting. For construction of any structure near
by a water body or in between a water body and for determination of safe levels of construction to protect structure from
flood water, safe grade elevation is required.
In order to evaluate or estimate, mitigate and handle the floods, the present paper presents a methodology for assessment of flood
line to produce safe grade elevation by using River Analysis System made by Hydrologic Engineering Center (HEC-RAS)
software which is predominately used in the field of hydraulic analysis for floodplain delineation. The general parameter affecting
flood is runoff gauge, discharge, rainfall and land use as spatial data. This paper explains the use of the HEC-RAS for producing
the safe grade elevation for Mutha River from its origin at downstream side of Kadakwasla dam till Mahtre Bridge. It explains the
methodology to construct a table model and how to validate it. The methodology developed can be applied for regions if only
predominant factors affecting the flood in that region is consider, to decide the best economical safe grade elevation for the
building or structure near or on the river and would help in planning priorities prerequisites for managing flood efficiently.
Keywords: Safe Grade elevation, Parameters, Mutha River, Flood, spatial data, Zoning
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
DETERMINATION OF NET FLOWS INTO ALMATTI RESERVOIR FROM CWC GAUGE DATA AND RES...IAEME Publication
This paper presents the determination of net flows into Almatti reservoir from CWC data and reservoir data. From the study it can be concluded that the average flow in to Almatti will be 574.86 TMC, the maximum inflow will be 1196.8 TMC and the minimum flow will be 166.99 TMC. The flows in annual in deficit years may reduce by about 50 TMC but there is no variation in the good years in the good years as the storage effects will take care of this aspect during good years. It can be concluded that there will be reduction of flows in the June and July flows in the ultimate scenario except in very good years.
Floodplain Modelling Materials and MethodologyIDES Editor
A floodplain is the normally dry land area adjoining
river or stream that is inundated during flood events. The
most common reason for flooding could be overtopping of river
or stream due to heavy downfall. The floodplain carries flow
in excess of the river or stream capacity. Flood frequency and
flood water-surface elevations are the crucial components for
the evaluation of flood hazard. This paper presents the
methodology that incorporates advanced technologies for
hydrologic and hydraulic analyses that are needed to be carried
out to predict the flood water-surface elevations for any
ungaged watershed.
Determination of safe grade elevation by using hec ras case study mutha rivereSAT Journals
Abstract
Flood is a naturally occurring disastrous event causing damages, losses and destruction to property, life and environment.
Hundred millions of money are spent every year in flood control and flood forecasting. For construction of any structure near
by a water body or in between a water body and for determination of safe levels of construction to protect structure from
flood water, safe grade elevation is required.
In order to evaluate or estimate, mitigate and handle the floods, the present paper presents a methodology for assessment of flood
line to produce safe grade elevation by using River Analysis System made by Hydrologic Engineering Center (HEC-RAS)
software which is predominately used in the field of hydraulic analysis for floodplain delineation. The general parameter affecting
flood is runoff gauge, discharge, rainfall and land use as spatial data. This paper explains the use of the HEC-RAS for producing
the safe grade elevation for Mutha River from its origin at downstream side of Kadakwasla dam till Mahtre Bridge. It explains the
methodology to construct a table model and how to validate it. The methodology developed can be applied for regions if only
predominant factors affecting the flood in that region is consider, to decide the best economical safe grade elevation for the
building or structure near or on the river and would help in planning priorities prerequisites for managing flood efficiently.
Keywords: Safe Grade elevation, Parameters, Mutha River, Flood, spatial data, Zoning
APPLICATION OF 1-D HEC-RAS MODEL IN DESIGN OF CHANNELSAM Publications
Flood occurs at Surat city frequently due to sudden release of water from Ukai dam in river Tapi. At the
time of floods in river Tapi, Surat city and surrounding regions are most affected. The city has faced many floods
since long back. Major flood event occurred in the year 1883, 1884, 1942,1944,1945,1949, 1959, 1968, 1994, 1998,
2002, 2006, 2007 and 2012. The carrying capacity of river is approximately about 4.5 lakhs cusecs (12753 cumecs) at
present. In this, stability of a segment of lower reach approximately 6 km length of Tapi river between Weir cum
causeway and Sardar bridge is evaluated for its carrying capacity and stability in response to discharge and slopes
using HEC-RAS software for past flood data. The study reach consists of 24 cross-sections. The hydraulics model,
HEC-RAS is employed to evaluate flood conveyance performance and also uniform flow computation is carried out.
In the present study existing storm drains are not only marked but based on the HEC-RAS water surface elevation
computation for various flood discharges, need of flood gates on the storm drains are also assessed. The
recommendations are done based on this study either to increase height of bank or construct a retaining wall at
certain sections along the study reach. The present study also recommends installations of flood gates on all the storm
drain outlets which are without flood gates. The width of river in no case be encroached as sections are sensitive high
floods.
Abstract Urban watersheds produce an instantaneous response to rainfall. That results in stormwater runoff in excess of the capacity of drainage systems. The excess stormwater must be managed to prevent flooding and erosion of streams. Management can be achieved with the help of structural stormwater Best Management Practices (BMPs). Detention ponds is one such BMP commonly found in the Austin, TX, USA. The City of Austin developed a plan to mitigate future events of flooding and erosion, resulting in the development and integration of stormwater BMP algorithms into the sub-hourly version of SWAT model. This paper deals with the development of a physically based algorithm for detention pond. The algorithm was tested using a previously flow-calibrated watershed in the Austin area. From the test results obtained it appears that the detention pond algorithm is functioning satisfactorily. The algorithm developed could be used a) to evaluate the functionality of individual detention pond b) to analyze the benefits of such structures at watershed or higher scales and c) as design tool. Keywords: flooding, detention, urban, watershed, BMP, algorithm, stormwater, modeling
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Hydrological Calibration in the Mount Lofty Ranges using Source Paramenter Es...eWater
The catchments of the Mount Lofty Ranges (MLR) provide a crucial water resource for the rural and urban community of Adelaide. The Source Catchments model is one planning tool used to assess current and future water, sediment and nutrient yields from the catchment. The modelling is critical for future planning to ensure a safe and reliable water supply is maintained.
Linking PEST and Source Catchments for hydrology calibration was an efficient and repeatable method for calibration. Future work to improve the process could include calibrating with other rainfall runoff models such as Sacramento which better account for groundwater losses or explicit representation of groundwater in the model.
This project has focused solely on rainfall and runoff. It has not considered how external impacts (e.g. climate change) may change catchment hydrological characteristics or constituent characteristics such as EMC/DWC values. Future projects through the Goyder Water Research Institute aim to address some of these
issues. This project may produce data which will inform further work in this respect.
The study examined the characteristics of the Sumanpa stream’s Flow-Duration-Frequency Curve statistics for a period of 25years (1985-2009) and compared the 1990-1999 and 2000-2009 Flow-Duration-Curves. The high, low and mean Flow-Duration-Curves were also analysed. The discharge records were analysed to develop a general quantitative characterization of the stream’s flow variability. Streamflow data was generated from daily stage data using the rating curve model developed at the stream’s gauge station. Flow-Duration-Frequency-Curves were developed using the Weibull plotting position and used to analyse the catchment’s surface and groundwater storage and stream’s flow characteristics. The approach placed the midpoints of the moist, mid-range, and dry zones of the curves at 25th, 50th, and 75th percentiles, respectively. The high zone was centered at the 5th percentile, while the low zone was centered at the 95th percentile. For 95% of the time, the streamflowequalled or exceeded 0.14 m3s-1, at 5% it equalled or exceeded 45 m3s-1 and at 50% flow equalled or exceeded 5.53 m3s-1.
This paper elaborates the hydraulic characteristics of the water supply network of the town of Puerto Ayora. First, it intends to replicate the household individual storage by simulating nodal tanks with the use of the EPANET software. Later, it uses the Pressure-Driven Approach (PDA) to develop a methodology that estimates the overflow of storage facilities, one of the main sources of wastage in Puerto Ayora. Finally, it uses the Demand-Driven Approach (DDA), with the aim of assessing the network in the future, under four population growth scenarios. With the chosen moderate growth scenario, two options are suggested in order to tackle the water supply issues at the end of the planning horizon.
APPLICATION OF 1-D HEC-RAS MODEL IN DESIGN OF CHANNELSAM Publications
Flood occurs at Surat city frequently due to sudden release of water from Ukai dam in river Tapi. At the
time of floods in river Tapi, Surat city and surrounding regions are most affected. The city has faced many floods
since long back. Major flood event occurred in the year 1883, 1884, 1942,1944,1945,1949, 1959, 1968, 1994, 1998,
2002, 2006, 2007 and 2012. The carrying capacity of river is approximately about 4.5 lakhs cusecs (12753 cumecs) at
present. In this, stability of a segment of lower reach approximately 6 km length of Tapi river between Weir cum
causeway and Sardar bridge is evaluated for its carrying capacity and stability in response to discharge and slopes
using HEC-RAS software for past flood data. The study reach consists of 24 cross-sections. The hydraulics model,
HEC-RAS is employed to evaluate flood conveyance performance and also uniform flow computation is carried out.
In the present study existing storm drains are not only marked but based on the HEC-RAS water surface elevation
computation for various flood discharges, need of flood gates on the storm drains are also assessed. The
recommendations are done based on this study either to increase height of bank or construct a retaining wall at
certain sections along the study reach. The present study also recommends installations of flood gates on all the storm
drain outlets which are without flood gates. The width of river in no case be encroached as sections are sensitive high
floods.
Abstract Urban watersheds produce an instantaneous response to rainfall. That results in stormwater runoff in excess of the capacity of drainage systems. The excess stormwater must be managed to prevent flooding and erosion of streams. Management can be achieved with the help of structural stormwater Best Management Practices (BMPs). Detention ponds is one such BMP commonly found in the Austin, TX, USA. The City of Austin developed a plan to mitigate future events of flooding and erosion, resulting in the development and integration of stormwater BMP algorithms into the sub-hourly version of SWAT model. This paper deals with the development of a physically based algorithm for detention pond. The algorithm was tested using a previously flow-calibrated watershed in the Austin area. From the test results obtained it appears that the detention pond algorithm is functioning satisfactorily. The algorithm developed could be used a) to evaluate the functionality of individual detention pond b) to analyze the benefits of such structures at watershed or higher scales and c) as design tool. Keywords: flooding, detention, urban, watershed, BMP, algorithm, stormwater, modeling
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Hydrological Calibration in the Mount Lofty Ranges using Source Paramenter Es...eWater
The catchments of the Mount Lofty Ranges (MLR) provide a crucial water resource for the rural and urban community of Adelaide. The Source Catchments model is one planning tool used to assess current and future water, sediment and nutrient yields from the catchment. The modelling is critical for future planning to ensure a safe and reliable water supply is maintained.
Linking PEST and Source Catchments for hydrology calibration was an efficient and repeatable method for calibration. Future work to improve the process could include calibrating with other rainfall runoff models such as Sacramento which better account for groundwater losses or explicit representation of groundwater in the model.
This project has focused solely on rainfall and runoff. It has not considered how external impacts (e.g. climate change) may change catchment hydrological characteristics or constituent characteristics such as EMC/DWC values. Future projects through the Goyder Water Research Institute aim to address some of these
issues. This project may produce data which will inform further work in this respect.
The study examined the characteristics of the Sumanpa stream’s Flow-Duration-Frequency Curve statistics for a period of 25years (1985-2009) and compared the 1990-1999 and 2000-2009 Flow-Duration-Curves. The high, low and mean Flow-Duration-Curves were also analysed. The discharge records were analysed to develop a general quantitative characterization of the stream’s flow variability. Streamflow data was generated from daily stage data using the rating curve model developed at the stream’s gauge station. Flow-Duration-Frequency-Curves were developed using the Weibull plotting position and used to analyse the catchment’s surface and groundwater storage and stream’s flow characteristics. The approach placed the midpoints of the moist, mid-range, and dry zones of the curves at 25th, 50th, and 75th percentiles, respectively. The high zone was centered at the 5th percentile, while the low zone was centered at the 95th percentile. For 95% of the time, the streamflowequalled or exceeded 0.14 m3s-1, at 5% it equalled or exceeded 45 m3s-1 and at 50% flow equalled or exceeded 5.53 m3s-1.
This paper elaborates the hydraulic characteristics of the water supply network of the town of Puerto Ayora. First, it intends to replicate the household individual storage by simulating nodal tanks with the use of the EPANET software. Later, it uses the Pressure-Driven Approach (PDA) to develop a methodology that estimates the overflow of storage facilities, one of the main sources of wastage in Puerto Ayora. Finally, it uses the Demand-Driven Approach (DDA), with the aim of assessing the network in the future, under four population growth scenarios. With the chosen moderate growth scenario, two options are suggested in order to tackle the water supply issues at the end of the planning horizon.
An immersive workshop at General Assembly, SF. I typically teach this workshop at General Assembly, San Francisco. To see a list of my upcoming classes, visit https://generalassemb.ly/instructors/seth-familian/4813
I also teach this workshop as a private lunch-and-learn or half-day immersive session for corporate clients. To learn more about pricing and availability, please contact me at http://familian1.com
3 Things Every Sales Team Needs to Be Thinking About in 2017Drift
Thinking about your sales team's goals for 2017? Drift's VP of Sales shares 3 things you can do to improve conversion rates and drive more revenue.
Read the full story on the Drift blog here: http://blog.drift.com/sales-team-tips
How to Become a Thought Leader in Your NicheLeslie Samuel
Are bloggers thought leaders? Here are some tips on how you can become one. Provide great value, put awesome content out there on a regular basis, and help others.
The Development of a Catchment Management Modelling System for the Googong Re...GavanThomas
A scenario assessment model to assist the end-user in determining priorities for a series of agreed management prescriptions that can be enacted through controls on existing landuse
A Holistic Approach for Determining the Characteristic Flow on Kangsabati Cat...ijceronline
Kangsabati river rises from the Chotanagpur plateau in the state of West Bengal, India and passes through the districts of Purulia, Bankura and Paschim Medinipur in West Bengal before joining into river Rupnarayan. It is life of these three districts of West Bengal situated in the western part of the state. The river has ephemeral characteristics i.e. it has low flow in the year round and have a high peak on a certain time basis. In the Kangasabati catchment hydrological study gives an evident that during the period every two years there is a chance of drought condition and consecutively after that there is a high flow year. In our study period from 1991 to 2010 there are six low streamflow year i.e. in that year there is less rainfall than the average rainfall on that area. The year 1991, 2002 and 2009 are the drought prone year and above that in 2010 the severe drought condition was seen and this is the lowest rainfall year among the last 20 years and the rainfall on this year is only 766 mm which is in an about 38% less rainfall than the average rainfall of the catchment. And the highest flood peak in the last twenty year is noted on 19th Aug 2007 as 377107.8 Mm3
Fitting Probability Distribution Functions To Discharge Variability Of Kaduna...IJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
Guidelines for Modelling Groundwater Surface Water Interaction in eWater SourceeWater
One of the key challenges in modelling GW-SW interactions is the significant time-scale
differences between surface water and groundwater processes. Because groundwater
movement can be orders of magnitude slower than surface water movement, the
responses of groundwater systems to hydrological and management drivers such as
climate variability, land use change, and groundwater extraction can be very damped and
lagged. Hence, a key requirement in modelling GW-SW interactions in river system
models is to account for these time lags.
The modelling of GW-SW interactions in river system models is still very much in its
infancy, not just in Australia, but also throughout the world. As such, there is no consensus
on implementation of this functionality in river system models, and hence the little
discussion in the literature so far on what constitutes Best Practice Modelling in this
domain.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Welcome to the first live UiPath Community Day Dubai! Join us for this unique occasion to meet our local and global UiPath Community and leaders. You will get a full view of the MEA region's automation landscape and the AI Powered automation technology capabilities of UiPath. Also, hosted by our local partners Marc Ellis, you will enjoy a half-day packed with industry insights and automation peers networking.
📕 Curious on our agenda? Wait no more!
10:00 Welcome note - UiPath Community in Dubai
Lovely Sinha, UiPath Community Chapter Leader, UiPath MVPx3, Hyper-automation Consultant, First Abu Dhabi Bank
10:20 A UiPath cross-region MEA overview
Ashraf El Zarka, VP and Managing Director MEA, UiPath
10:35: Customer Success Journey
Deepthi Deepak, Head of Intelligent Automation CoE, First Abu Dhabi Bank
11:15 The UiPath approach to GenAI with our three principles: improve accuracy, supercharge productivity, and automate more
Boris Krumrey, Global VP, Automation Innovation, UiPath
12:15 To discover how Marc Ellis leverages tech-driven solutions in recruitment and managed services.
Brendan Lingam, Director of Sales and Business Development, Marc Ellis
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Enhancing Performance with Globus and the Science DMZGlobus
ESnet has led the way in helping national facilities—and many other institutions in the research community—configure Science DMZs and troubleshoot network issues to maximize data transfer performance. In this talk we will present a summary of approaches and tips for getting the most out of your network infrastructure using Globus Connect Server.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Elevating Tactical DDD Patterns Through Object Calisthenics
Q4103103110
1. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
RESEARCH ARTICLE
www.ijera.com
OPEN ACCESS
Prediction of River Stage in an Ungauged Stream
*Onosakponome O. R, ** Mbachu V. C and ***Odenigbo C
*Department of Civil engineering, Madonna University, Akpugo campus.
**Department of Civil engineering, Madonna University, Akpugo campus
***Department of Civil engineering, Enugu State University of Science and Technology.
ABSTRACT
Hydrologic decisions are made to evaluate values of key variables and parameters needed for the design of any
water resources system to make it perform adequately – in terms of safety and provision of the expected
benefits. However, the degree of reliability of the hydrologic data used in the decision making process is of
great significance. Unreliable data will seriously affect results. In Nigeria, as well as many parts of the less
developed world, there are problems of data inadequacy, frequent gaps in the data available and non-existent
data at development sites. This is the dilemma that confronts any designer of water resources systems in the less
developed world. There is no doubt that sufficient and accurate hydrological data will lead to a sound
engineering design of the water resources system. This study presents a solution for prediction of river stage in
ungauged stream using the Principal Component Analysis (PCA). The research was illustrated by using the Imo
River with a station at OBIGBO as a case study, upon which data was collected for analysis and possible
development of a model. Twelve input variables were considered in the analysis; the most important
contribution of PCA in this study was the identification of the key factors responsible for the changes in river
stage. The amount of precipitation and the run-off discharge into the stream were the factors identified by the
PCA, which can reasonably reflect the status of river stage in many streams. The developed model did not only
predict the river stage of OBIGBO but also show great level of accuracy in predicting that of NEKEDE with an
average correlation coefficient of 0.95. It can be concluded that the model has a great ability to predict river
stage.
Keywords- Design, Hydrologic data, Prediction, River stage Water resources.
I.
INTRODUCTION
River stage or flow rates are required for the
design and evaluation of hydraulic structures. Most
river reaches are ungauged and a methodology is
needed to estimate the stages, or rates of flow, at
specific location in streams where no measurements
are available. Flood routing techniques are utilized to
estimate the stages, or rates of flow, in order to
predict flood wave propagation along river reaches.
Models can be developed for gauged catchments and
their parameters related to physical characteristics
such as slope, reach width, reach length so that the
approach can be applied to ungauged catchments in
the region.
The design, planning, and operation of river
systems depend largely on relevant information
derived from the forecasting and estimation of
extreme events. Reliable flood forecasts are
particularly important for improving public safety
and mitigating economic damages caused by
inundations. During the past few decades, a great
deal of research has been devoted to the modeling
and forecasting of river flow dynamics. Such efforts
have led to the formulation of a wide variety of
approaches and the development of a large number of
models. The existing models for river stage
www.ijera.com
forecasting may broadly be grouped under two main
categories namely, rainfall-runoff modeling or
statistical techniques. Due to the realistic
representation of watershed topography and ability to
capture the surface and ground water interaction, the
more reasonable method to predict a flood is the
distributed and physically based model. However,
extensive
topographic,
meteorological,
and
hydrologic data are required to describe the runoff
process and time is also require to calibrate
conceptual models (especially distributed models),
which are important factors to be considered in their
practical applications. Thus, the implementation and
calibration of conceptual models can typically
present various difficulties (Hu and Lam, 2001). In
this context data-driven models, which can discover
relationships from input-output data without having
the complete physical understanding of the system,
may be preferable. While such models do not provide
any information on the physics of the hydrologic
processes, they are in particular, very useful for river
flood forecasting where the main concern is accurate
prediction of a flood at specific watershed locations
(Nayak, 2005). Flooding is a type of natural disaster
that has been occurring for centuries, but can only be
mitigated rather than completely solved. Prediction of
103 | P a g e
2. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
river stages becomes an important research topic in
hydrologic engineering. An accurate water stage
prediction allows the pertinent authority to issue a
forewarning of the impending flood and to implement
early evacuation measures when required.
II.
THE ESTIMATION, PREDICTION
AND FORECASTING OF RUNOFF
Ideally, all hydrological problems would be
solved by the use of measured data, thus obviating
the necessity for estimation, prediction, and
forecasting. There are many circumstances, however,
in which the use of these techniques becomes
necessary. Thus, for example, there may be a
deficiency of measured data for a particular area, but
there may be the possibility of extrapolating future
runoff trends either from existing runoff data relating
to adjacent or nearby areas or from precipitation data.
Alternatively, measured data may be collected too
late to be of any use. Such is the case in areas where
peaks of quickflow constitute a flood problem which
must be viewed and solved in the light, not only of
hydrological factors, but also of factors of settlement
and communications, agriculture, and economics.
Inevitably, the relevant measured data cannot become
available until the flood peaks themselves have
occurred and so, in these circumstances, the need is
for techniques for accurately forecasting the volume
and timing of quickflow peaks (Mesfin, 2008).
Again, in areas where water supplies for agriculture,
industry, or domestic uses are likely to be limited at
times of low flow; the need is for accurate forecasts
of the magnitude of dry-weather flows, and the time
occurrence of minimum flow.[2-8]
The main requirements, therefore; are for
techniques to forecast, for a given point within a
drainage basin, both the total volume of runoff and
the magnitude of the instantaneous peaks normally
associated with sudden increases of quickflow and
also to forecast the timing and magnitude of the
minimum flows which are likely to be associated
with decreasing volumes of baseflow, particularly
groundwater flow. Most of the techniques currently
in use were developed before the newer concepts of
runoff formation. Interestingly, however, many of
these methods yield reasonable results despite being
conceptually weak or even erroneous. Such successes
may be fortuitous but techniques are either highly
empirical, and are often applicable only to restricted
areas, or else are based upon factors which, although
not directly cause-related to the patterns of runoff
under consideration, are themselves directly affected
by the real runoff-forming factors.
Although, in normal English usage, the
terms forecasting and prediction are clearly
synonymous, they are sometimes used in a more
restricted sense by hydrologists. Thus, as Smith
www.ijera.com
www.ijera.com
(1972) observed, prediction, in this context, refers to
the application of statistical concepts to long periods
of data, usually relating to extreme events, with a
view to defining the statistical probability or return
period of a given magnitude of flow. In other words,
there is no indication of when this particular flow will
occur. Forecasting, on the other hand, refers to
specific runoff events, whether floods or low flows,
and to the use of current hydro-meteorological data in
order to provide a forecast of the magnitude of the
runoff event and also, in many cases, of its timing. As
far as possible, this distinction will be preserved in
the ensuring discussions.
There are many techniques of runoff
prediction and forecasting. Some of these are in
widespread use, either because they work reasonably
well over a wide range of conditions or else are easy
to apply. The use of other techniques may be
restricted to specific areas or to specific users, such
as a particular Government agency. Most methods
have little merit and yield poor results. It would
clearly be impossible to deal with all methods or even
a representative selection of the better ones. Indeed,
in the present context this would not, in any case, be
appropriate. Instead, the main lines of approach to the
problem of river stage or runoff prediction and
forecasting will be briefly reviewed in general terms
and will be illustrated, where appropriate, by specific
examples.
2.1 THEORY OF PRINCIPAL COMPONENT
ANALYSIS
Principal component analysis (PCA) is a
statistical technique for determining the key variables
in a multidimensional data set that explains the
differences in the observations and can be used to
simplify the analysis and visualization of
multidimensional data sets. In recent years, the
method of principal component analysis has been
widely used in many fields such as evaluation of
irrigation water quality, evaluation of river water
quality monitoring stations and comprehensive
evaluation of the regional water resource carrying
capacity.
The method of principal component analysis
(PCA), using coefficients of linear correlation offers
this possibility. Principal component analysis is also
known as eigenvector analysis, eigenvector
decomposition. Generally, the principal component
of random vector X is obtained from the weight of X
by special linear combination. Therefore, it is
difficult to give explanation for the physical meaning
of this linear combination when the dimensionless
variables are different. In order to perform a PCA of
the original data, random variables X have to be
standardized. PCA seeks to establish combinations of
variables capable of describing the principal
104 | P a g e
3. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
tendencies observed while studying a given matrix.
In mathematical terms, PCA relies upon an
eigenvector decomposition of the covariance or
correlation matrix. The basic ideas of principal
component analysis are to define Fn as a linear
combination of weight X, find a linear combination
of Fn for the weight X, and Fn reflects the changes of
the weight X as far as possible. Here, F1 is called the
first principle component of X, and if it not yet fully
reflects the changes of the weight X, we then find F2,
F3,…..,Fr(r<n), till the information of the weight X
was fully extracted and F can be given as:
Fn=A1X1+A2X2+...+AnXn
(1.1)
Given X observations on n variables, the
goal of PCA is to reduce the dimensionality of the
data matrix by finding r new variables, where r is less
than n. Each principal component is a linear
combination of the original variables, and so it is
often possible to ascribe meaning to what the
components represent.
It is used when there are a large number of
different types of measurement for a given set of
items. It aims at structuring the data by reducing the
numerous variables to a smaller number of variables
(components) which account for most of the variation
in given data. It is a powerful technique which copes
with the problems in both linear and non-linear least
squares associated with statistical interrelations
amongst the independent variables. It transforms the
independent variables into new variables that are
statistically unrelated.
The
principal
component
analysis
transforms the linear model
Q = c1 x1 +… +cpxp
Where
1 is a component or eigenvector
(1.4)
(i.e. the new variable are statistically
independent)
The problem is made simpler by removing the scale
effects of the original variables. Hence, the
normalized original variables are defined.
x1 x
s1
n
Zu 0
j 1
And
n
rik=
u
Z kj
(Mean)
(1.6)
(1.8)
3
3
3
3
3
3
3
3
3
Fig.1.1: Location of components in three dimensions
The simple correlation matrix with
eigenvalue in the diagonal is given as
1
r21
rp 2
r12 ....r1 p
1 .......r2 p
r 2
p .........
1
L1
L
2=0
L p
as the
(1.9)
Where L1------Lp represent the direction cosines i.e.
the cosine of the angles of rotation of the axis.
Equation (1.9) can be written as
r 1 l 0
(2.0)
Since the directional cosine vector i.e. L =
(Li………..L1) is nonzero
r 1 0
(2.1)
www.ijera.com
Z
j 1
(1.5)
Letting the mean be zero and the standard deviation
be unity. The first two moments of the normalized
variants are
(1.7)
The solution to our problem begins by
“plotting” all p of the original variables in pdimensional space and rotating the axis until the
orthogonal system of components is found. An
attempt to demonstrate this for three dimensions is
shown in fig.1.1. The data points are plotted as
referenced to all the axes of the three original
variables and then the axes are rotated until the
components are orthogonal or statistically
independent.
Statistically, this feat is achieved by
minimizing the variance or spread around the
components subject to the constraint that
orthogonality must be achieved.
The simple correlation coefficient rik is
given as
(1.3)
such that
Cov { , } ,
Z1=
1 n 2
Z u 1 (Standard deviation)
n j 1
(1.2)
To Q = β1ε1 +…+ βpεp
www.ijera.com
105 | P a g e
4. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
Equation (2.1) is solved by expanding it as a
determinant and obtaining an algebraic equation to
the pth power.
C1 P C 2 P 1 ... Cp 1 0
(2.2)
Equation (2.2) has p-roots, the eigenvalues.
With each of these values, the eigenvectors are found
by substituting back each of the eigenvalues in
equation.
The principal components are found by
noting the following; the variance of Za is
V
Z a V li
(2.3)
Where,
1
2
V Z L11 1 L21 2 ... L p1 p
n
(2.4)
This result in
V
2
Z L11 L2 ... L2pl
21
Or
Hence,
V Z
(2.5)
p
L i
2
(2.6)
i 1
Li is a type of correlation coefficient
representing the correlation of Z p with . The
III.
DATA AND METHOD
In this study, the data were mainly collected
from the Anambra-Imo River Basin Development
Authority; statistical year book 1999-2009. In the
development of a model for predicting the water
stage in River Imo with station at OBIGBO, 12
variables were considered believed to be contributing
significantly to the water level fluctuation using the
Principal Component Analysis (PCA).
The 12 variables were;
(1)
Evaporation rate (X1)
(2)
Stream Discharge (X2)
(3)
Stream width (X3)
(4)
Average velocity of flow (X4)
(5)
Channel slope (X5)
(6)
Infiltration rate (X6)
(7)
Runoff discharge into the stream (X7)
(8)
Population size (X8)
(9)
Efficiency of drainage network (X9)
(10) Catchment area (X10)
(11) Precipitation amount (X11)
(12) Length of main stream (X12)
The simple correlation matrix of the 12
variables was obtained. There are several significant
correlations of the variables (Xi) with the water
stage(y) but at a glance it would be nearly impossible
to decide which one to choose. Further, there are a
number of significant interrelations amongst the
independent variables, but since the entire correlation
procedure is a matter of degree, it would be
impossible to filter out objectively the variables at
www.ijera.com
www.ijera.com
directional cosine squared is a variance, which
represents the fraction of Z explained by .
Once principal component analysis has been
completed, the regression can be performed whereby
is considered to be the independent variables.
Then, the model coefficient Ci in equation (1.2) can
be derived from the regression coefficients in
equation (1.3).
Equating the two models results in
L11 L21 ...L p1
L12 L22 ...L p 2
L1 p L2 p[ ...L pp
C 1
1
C
2 = 2
C p
p
(2.7)
And the solution of Ci is
C = L-1
(2.8)
When component analysis is combined with
nonlinear least squares, the original variables are
Q
and the coefficients are hi.
j
this point. In addition, the standardized values of
factors were calculated from the original data by
SPSS.
3.1 Model performance
The performance of the model developed in
this study was assessed using various standard
statistical performance evaluation criteria. The
statistical measures considered were multiple
correlation coefficients (MCC), standard error of
estimate (SEE), coefficient of correlation (CORR),
mean absolute percentage error (MAPE), and root
mean square error (RMSE).
IV.
DATA ANALYSIS
Table 1.0: Percent variance of the 12 variables
Variables
X1
I
82.17
II
12.05
III
4.03
Total
98.25
X2
X3
X4
87.59
36.64
2.99
5.56
18.98
42.59
3.97
37.12
0.08
97.12
92.74
45.66
X5
X6
X7
X8
4.60
0.30
1.28
42.11
67.16
83.01
1.03
26.70
5.90
0.64
84.05
29.01
77.66
83.95
86.36
97.82
X9
0.07
31.20
54.95
86.22
X10
87.97
4.91
6.93
99.81
X11
92.49
4.99
0.56
98.04
X12
70.02
9.97
6.04
86.03
106 | P a g e
5. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
Table 1.1: Eigenvalue contribution rates and
accumulated contribution rates of the principal
components.
Component Eigenvalue %
of Cumulative
variance %
1
5.08
42.33
42.33
2
3.08
25.67
68.00
3
2.33
19.42
87.42
The principal component analysis (PCA)
operates as a filter of redundant information and as
mechanism for model building. The equation, (A-λI)
= 0 was solved, where A is the correlation matrix of
variables. Because it led to a 12 x 12 determinant, it
was solved by computer. The first three largest
eigenvalues (i.e., the values of λ) were selected i.e.,
5.08, 3.08 and 2.33 in descending order and the
others rejected. With each of these values, the
eigenvectors were obtained by substituting each at a
time in the equation [A-λI] [L] =0 (equation 1.9). The
eigenvectors were normalized. We see that
component 1 is highly correlated with X1, X2, X10,
X11 and X12 indicating that these five variables are
highly interrelated since eigenvectors are made like
correlation coefficients. In component 2, we see that
it is highly correlated with X5 and X6, while
component 3 is highly correlated with X7 and X9
indicating that these two variables are highly
interrelated. The variance are explained by the
components which are the eigenvectors squared and
are referred to as loading.
Our analysis indicates that we can
summarize the data with just three components.
Table 1.1 contains the three principal components
and their corresponding eigenvalues. The results
showed that of the first three components, the first
component accounted for about 42.33%, the second
component about 25.67% and the third component
about 19.42% of the total variance in the data set.
These three components together accounted for about
87.42% of the total variance and the rest of the
components only accounted for about 12.58%.
Therefore, our discussion focused only on the first
three components.
The following decisions were made; only
variable X11 was used since it explains 92.49% of the
information contained in the component and is by far
the easiest of the five interrelated variables to
measure.
Only variable X7 was used since it explains
84.05% of the information contained in that
component.
The remaining components were deleted because
the marginal variance that they explain was
deemed insignificant.
Upon making the above decision, the
formulated model was;
www.ijera.com
www.ijera.com
Y ax b x c
m
7 11
(2.9)
Where Ym is mean annual river stage and, a, b and c
are constants.
The model was transformed to linear form by taking
the natural log of equation (2.9) to get:
In Ym = ln a + bln X7 + cln X11
(3.0)
This reduces to the form;
Y = z + mQ + nP
(3.1)
Where P and Q are the precipitation amount
(mm) and runoff discharge (m3/s) into the stream
respectively, z, m and n are regression constants.
Using the highest values of the data set (1999-2005),
the constants were determined using the least square
method. The model is given by;
Y=4.02+0.009P+0.014Q
(3.2)
The multiple correlation coefficient, MCC = 0.95
while the standard error of estimate, SEE = 0.2
4.1
Results and Discussion
The available data set was divided into two
sets, from (1999 – 2005) and (2006 – 2009). The first
data set was used to perform the principal component
analysis and to calibrate the resulting model. The
second data set was used for verification. During the
verification phase, attempt was made for the
validation of the model by its application to predict
the river stage of OTAMIRI with its station in
NEKEDE. It has a catchment area of 100 SQ KM and
is located within the Imo river system. The following
statistical parameters were used for the evaluation
and the results presented in table 1.2 below. The
parameters are; MAPE, it measures the absolute error
as a percentage of the forecast, and RMSE evaluates
the residual between observed and predicted river
stage. CORR evaluates the linear correlation between
the observed and predicted river stage.
Table 1.2; Performance of the
stations
Station Performance
OBIGBO
Year CORR
MAPE
RMSE
2006 0.95
2.15
7.49
2007 0.97
1.34
3.12
2008 0.98
1.15
2.65
2009 0.96
2.56
4.11
model at different
NEKEDE
CORR
RMSE
0.94
8.00
0.96
3.55
0.96
5.02
0.95
5.51
MAPE
2.34
2.15
1.73
3.78
From the result in 2006, the model
performance at OBIGBO in terms of CORR, MAPE
and RMSE were 0.95, 2.15 and 7.49, respectively,
which were better than those obtained at NEKEDE
107 | P a g e
6. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
(0.94, 2.34 and 8.00 respectively). The same result
applies to the remaining years (2007, 2008 and
2009), the reason being that the model was developed
using OBIGBO data set. However, the model proved
valid being able to predict to high degree of accuracy,
the river stage of OTAMIRI with station at
www.ijera.com
NEKEDE. See graphical illustrations in fig.1.2-1.9
below.
River Stage (m)
8
Predicted
6
Measured
4
2
0
Jan
Feb
Mar
Apr
May
Jun
Jul
Period (Month)
Aug
Sep
Oct
Nov
Dec
Fig. 1.3: Comparison between measured and predicted River Stage in
Obigbo, 2007
Predicted
Measured
6
4
2
0
Jan
Feb
Mar
Apr
May
Jun
Jul
Period (month)
Aug
Sep
Oct
Nov
Dec
Fig 1.4: Comparison between measured and predicted River Stage in Obigbo, 2008.
Predicted
Measured
River Stage (m)
10
8
6
4
2
0
Jan
Feb
Mar
Apr
May
Jun
Jul
Period (month)
Aug
Sep
Oct
Nov
Dec
Fig 1.5: Comparison between Measured and Predicted River Stage in Obigbo, 2009
River Stage (m)
River Stage (m)
8
8
7
6
5
4
3
2
1
0
Predicted
Measured
Jan
Jun
Jul
Aug
Sep
Oct
Period (month)
Fig. 1.6: Comparison between Measured and Predicted River Stage in Nekede, 2006
www.ijera.com
Feb
Mar
Apr
May
Nov
108 | P a g e
Dec
7. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
V.
CONCLUSION
This study presents a solution for prediction
of river stage in ungauged stream using the Principal
Component Analysis (PCA). The research was
illustrated by using the Imo River with a station at
OBIGBO as a case study, upon which data was
collected for analysis and possible development of a
model. Twelve input variables were considered in the
analysis; the most important contribution of PCA in
this study was the identification of the key factors
responsible for the changes in river stage. The
amount of precipitation and the run-off discharge into
the stream were the factors identified by the PCA,
which can reasonably reflect the status of river stage
in many streams. The developed model did not only
predict the river stage of OBIGBO but also show
great level of accuracy in predicting that of NEKEDE
www.ijera.com
www.ijera.com
with an average correlation coefficient of 0.95. It can
be concluded that the model has a great ability to
predict river stage in a homogenous catchments and
the predicted results provide a useful guidance or
reference for flood control operations. Yet, more
substantial improvement certainly should be pursued
through further research to improve the forecast
results.
REFERENCES
[1]
[2]
Ahsan, M. & K.M. O’Connor (1994). A
simple non-linear rainfall-runoff model with
a variable gain factor. J. Hydrol., 155. pp.
151 – 183.
Ayub, S., et al (2005). “Flood and drought
forecasting and early warning program (for
the Nile Basin)”. Nile Basin Capacity
109 | P a g e
8. Mbachu V. C et al Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 1( Version 3), January 2014, pp.103-110
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
www.ijera.com
Building Network for River Engineering
(NBCBN - RE).
Bengraine, K and Marhaba, T.F. 2003.
Using principal component analysis to
monitor spatial and temporal changes in
water quality. Journal of Hazardous
Material, 3: 179 – 195
Betson, R.P., (1964) “What is watershed
runoff?” JGR, 69: 1541 – 1551.
Chang, F.J. & Chen, Y.C. (2001) A
counterpropagation fuzzy-neural network
modeling approach to real time streamflow
prediction. J.Hydrol., 245, 153 – 164.
CHOW, V.T. (1964) “Runoff” section 14 in
Chow, V.T. (ed.). Handbook of Applied
Hydrology, McGraw_Hill, New York.
Clarke, R.T. (1994). Statistical Modeling in
Hydrology. John Wiley and Sons.
Cunge, J.A. (1969). “On the subject of a
flood propagation computation method
(Muskingum method), Journal of Hydraulic
Research, 7 (2).
Dalrymple, T. (1960) “Flood frequency
analysis”: Manual of Hydrology, part 3,
flood flow techniques, USGS Wat. Sup.
Pap.,1543-A 80 pp.
Dawdy, D.R. & O’Donnell, T., (1965).
Mathematical Models of Catchment
Behaviour. J. of the Hydraulics Div.,
A.S.C.E., Vol. 91, Number HY4 Proc. Paper
4410. pp. 123 – 137.
De Roo, A & G. Schmuck (2003). European
Flood Alert System. Joint Research Center,
Institute for Environment and Sustainability.
European Commissions.
De Zhou, Rougquu, Z., Liming, L.,
Linghing, G. and Simin, C. (2009). “A study
on water resources consumption by principal
component
analysis
in
Qingtongxia
irrigation areas of Yinchuau plain, China”.
Journal
of
food,
Agriculture
and
environmental Vol. 7. (3 & 4): 734 – 738.
Elzein, A.S. & I.S. Adam (1996). Sudan
Flood Early Warning System (FEWS).
ElMoutandis, Vol. 2, No. 5, pp. 23 – 24.
Hu, T.S., Lam, K.C. & Ng. S.T. (2001)
River flow time series prediction with a
range-dependent neural network. Hydrol.
Sci. J., 46, 729 – 745.
Mesfin, H. T (2008). “Flood routing in
ungauged catchments using Muskingum
Methods”. Dissertation.com, Boca Raton,
Florida, USA.
www.ijera.com
110 | P a g e