How software size influence productivity and project durationIJECEIAES
To make a perfect project plan, the software size of the order from the customer is the most important factor. The biggest challenge for the project manager is to estimate the project end date in the beginning of the project i.e. in project planning phase with realistic accuracy. Apart from other major inputs to estimate the project end date, expected team capability (productivity) and estimated software size are the major inputs that may influence the project end date. Software size is one of the most significant independent metric available in the planning phase and project manager has to estimate the other metrics based on the initial estimated software size. There is no direct relationship available between software size and project duration or software size and team productivity, however, there are industry data published by Quantitative Software Management and ISBSG that shows how these metrics influence each other. In this paper, using the data published by ISBSG and Quantitative Software Management, we try to statistically establish how productivity and project duration are influenced by software size. We have done linear regression analysis by generating the secondary data based on the data published by ISBSG and Quantitative Software Management. Linear regression equation validated with the actual project data and experimental results suggest that that productivity is significantly dependent on software size, however, project duration does not significantly depend on software size but may also be dependent on other metrics like team size, apart from software size.
Using The Advancement Degree Of Difficulty (Ad2) As An Input To Risk Managementjbci
Presentation describes the use of the Advancement Degree of Difficulty (AD2) as a means of identifying risk, cost and schedule associated with maturing technologies.
Why is TDD so hard for Data Engineering and Analytics Projects?Phil Watt
This slide show describes the difficulties in implementing Test-Driven Development (TDD) in the context of analytics and data engineering in development and maintenance phases. If we assumes that the objective of TDD is to reduce cycle time, improve developer productivity and improve production quality. It identifies 7 challenges from the analytics literature and a further 10 from interviews (n=14) and survey respondents (n=20) selected from analytics leaders. A key theme emerging as an output is that many of the challenges can be addressed through education and coaching, notably around data literacy for key stakeholders and executives
How software size influence productivity and project durationIJECEIAES
To make a perfect project plan, the software size of the order from the customer is the most important factor. The biggest challenge for the project manager is to estimate the project end date in the beginning of the project i.e. in project planning phase with realistic accuracy. Apart from other major inputs to estimate the project end date, expected team capability (productivity) and estimated software size are the major inputs that may influence the project end date. Software size is one of the most significant independent metric available in the planning phase and project manager has to estimate the other metrics based on the initial estimated software size. There is no direct relationship available between software size and project duration or software size and team productivity, however, there are industry data published by Quantitative Software Management and ISBSG that shows how these metrics influence each other. In this paper, using the data published by ISBSG and Quantitative Software Management, we try to statistically establish how productivity and project duration are influenced by software size. We have done linear regression analysis by generating the secondary data based on the data published by ISBSG and Quantitative Software Management. Linear regression equation validated with the actual project data and experimental results suggest that that productivity is significantly dependent on software size, however, project duration does not significantly depend on software size but may also be dependent on other metrics like team size, apart from software size.
Using The Advancement Degree Of Difficulty (Ad2) As An Input To Risk Managementjbci
Presentation describes the use of the Advancement Degree of Difficulty (AD2) as a means of identifying risk, cost and schedule associated with maturing technologies.
Why is TDD so hard for Data Engineering and Analytics Projects?Phil Watt
This slide show describes the difficulties in implementing Test-Driven Development (TDD) in the context of analytics and data engineering in development and maintenance phases. If we assumes that the objective of TDD is to reduce cycle time, improve developer productivity and improve production quality. It identifies 7 challenges from the analytics literature and a further 10 from interviews (n=14) and survey respondents (n=20) selected from analytics leaders. A key theme emerging as an output is that many of the challenges can be addressed through education and coaching, notably around data literacy for key stakeholders and executives
Why is Test Driven Development for Analytics or Data Projects so Hard?Phil Watt
Preview of research results for my Master's thesis on Test-Driven Development in Analytics. Prepared for my Term 4 assignment, oral thesis presentation
Risk Driven Approach to Test Device Softwareijtsrd
Software testing is one of the most crucial testings in the software development process. Software testing should be scheduled and managed very effectively. The risk is the situation that has not occurred yet and may not occur in the future as well. After looking at this definition, risks can refer to the probability of the failure for a particular project. Risk based testing is the type of testing that is based on the priority and importance of the software that has to be tested. In this research work, the new technique to test the device software has been proposed using the JAVA language. The new system is able to test the software based on various risks and provide alternatives based on that the risk can be reduced in the future. It also calculates the updated cost and duration required to complete the software when a risk has occurred. The proposed application is able to provide efficient and accurate results in terms of entered risks on the device software. In the future, the software can be used to test the device software for more number of risks to make it more suitable as per the user's requirements. Ashwani Kumar | Prince Sood "Risk Driven Approach to Test Device Software" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd25230.pdfPaper URL: https://www.ijtsrd.com/computer-science/other/25230/risk-driven-approach-to-test-device-software/ashwani-kumar
Fine–grained analysis and profiling of software bugs to facilitate waste iden...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Overview of Software Development Life Cycle Models. Why traditional parametric estimating tools do not help estimate a software project developed using the Agile model. Explain and demonstrate the “nearest neighbor” analogy technique to estimate Agile software projects. Data and actions needed to implement the nearest neighbor technique
Traditional project management methods are based on scientific principles considered “normal science,” but lack a theoretical basis for this approach. These principles make use of linear step–wise refinement of the project management processes using a planning–as–management paradigm. Plans made in this paradigm are adjusted by linear feedback methods. These plans cannot cope with the multiple interacting and continuously changing technology and market forces. They behave as a linear, deterministic, Closed–Loop control system.
ECONOMIC OUTPUT UNDER THE CONDITIONS OF SOCIAL FREEDOM IN SOFTWARE DEVELOPMENTijseajournal
Software developers organize their work autonomously. Agile development approaches give freedom to developers. However, the discussion about economy often leads to the comparison with manu-facturing processes which are tightly organised. If developers spend too much time on inefficient activities, the performance might decrease. The worst example could be breaks. Breaks could be seen as a reason for
economic problems by the management. This paper investigates the impact of breaks on the overall performance in software development. The investigations assume that if developers make pauses in a normal manner, this has not a negative impact on the profitability. In practice, the human-centric development process brings together a business process and a social process. The interaction of both processes was simulated. Due to the execution of 1.500 simulations, we obtained information on the economic progression of the development process under the influence of
breaks. We determined the impact of breaks on the overall profitability. This investigation contributes to the discussion of freedom in software development. It helps man-agers to
assess if employees who make breaks harm the profitability. This could lead to the imple-mentation of further business constraints.
Making decision in the presence of uncertainty requires estimating the impact of the outcome of those decisions. Here;s a collection of resources can can be used to guide that process
5.0 Estimating Agile Development ProjectsGlen Alleman
The primary purpose of software estimation is not to predict a project’s outcome; it is to determine whether a project’s targets are realistic enough to allow the project to be controlled to meet them ‒ Steve McConnell
SECURITY FOR DEVOPS DEPLOYMENT PROCESSES: DEFENSES, RISKS, RESEARCH DIRECTIONSijseajournal
DevOps is an emerging collection of software management practices intended to shorten time to market for
new software features and to reduce the risk of costly deployment errors. In this paper we examine the security implications of two of the key DevOps practices, automation of the deployment pipeline using a deployment toolchain and infrastructure-as-code to specify the environment of the eployed software. We focus on identifying what changes when an organization moves from manual deployments to DevOps
automated deployment processes.We reviewed the literature and conducted three case studies using simple configurations of common DevOps tools. This allowed us to identify specific:
• Positive influences on security where automation enhances defenses.
• Negative influences, where automation enables different kinds of attacks and increases the attack
surface.
• Research directions that look promising to support this new approach to software management.
• Recommendations for DevOps adopters
The primary purpose of software estimation is not to predict a project’s outcome; it is to determine whether a project’s targets are realistic enough to allow the project to be controlled to meet them ‒ Steve McConnell
Avoid software project horror stories - check the reality value of the estima...Harold van Heeringen
Many large software projects turn into software horror stories, resulting in newspaper headlines and even political issues. Often, the project costs and schedule were estimated unrealistically optimistic, using immature estimation techniques. A relatively simple way to avoid many problems is to perform a reality check on the estimate. This presentation was given on the conference of the International Cost Estimating and Analysis Association (ICEAA2014), June 2014 (Denver, USA)
Software Architecture Erosion: Impacts, Causes, and ManagementCSCJournals
As extensive new software projects are becoming more costly with increased development complexity and risks, the use of existing, already developed software projects are now spanning multiple years or decades. Software maintenance durations are increasing along with periodic periods of intense software upgrades due to these lengthy extensions of the life of the software. While the software architecture is set during the project's initial design phase, over the life of the project, architecture erosion can take place due to intentional and unintentional changes, resulting in deviations from the intended architecture. These deviations can manifest themselves in a wide variety of performance, maintenance, and software quality problems ranging from minor annoyances to product failure or unmaintainable software. This paper explores the causes and impacts of software erosion, and recovery approaches to addressing software erosion, laying the framework for future work towards the definition of an Architectural Maturity Model Integration process to assess the architectural maturity of an organization.
DevOps shifting software engineering strategy Value based perspectiveiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Why is Test Driven Development for Analytics or Data Projects so Hard?Phil Watt
Preview of research results for my Master's thesis on Test-Driven Development in Analytics. Prepared for my Term 4 assignment, oral thesis presentation
Risk Driven Approach to Test Device Softwareijtsrd
Software testing is one of the most crucial testings in the software development process. Software testing should be scheduled and managed very effectively. The risk is the situation that has not occurred yet and may not occur in the future as well. After looking at this definition, risks can refer to the probability of the failure for a particular project. Risk based testing is the type of testing that is based on the priority and importance of the software that has to be tested. In this research work, the new technique to test the device software has been proposed using the JAVA language. The new system is able to test the software based on various risks and provide alternatives based on that the risk can be reduced in the future. It also calculates the updated cost and duration required to complete the software when a risk has occurred. The proposed application is able to provide efficient and accurate results in terms of entered risks on the device software. In the future, the software can be used to test the device software for more number of risks to make it more suitable as per the user's requirements. Ashwani Kumar | Prince Sood "Risk Driven Approach to Test Device Software" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd25230.pdfPaper URL: https://www.ijtsrd.com/computer-science/other/25230/risk-driven-approach-to-test-device-software/ashwani-kumar
Fine–grained analysis and profiling of software bugs to facilitate waste iden...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Overview of Software Development Life Cycle Models. Why traditional parametric estimating tools do not help estimate a software project developed using the Agile model. Explain and demonstrate the “nearest neighbor” analogy technique to estimate Agile software projects. Data and actions needed to implement the nearest neighbor technique
Traditional project management methods are based on scientific principles considered “normal science,” but lack a theoretical basis for this approach. These principles make use of linear step–wise refinement of the project management processes using a planning–as–management paradigm. Plans made in this paradigm are adjusted by linear feedback methods. These plans cannot cope with the multiple interacting and continuously changing technology and market forces. They behave as a linear, deterministic, Closed–Loop control system.
ECONOMIC OUTPUT UNDER THE CONDITIONS OF SOCIAL FREEDOM IN SOFTWARE DEVELOPMENTijseajournal
Software developers organize their work autonomously. Agile development approaches give freedom to developers. However, the discussion about economy often leads to the comparison with manu-facturing processes which are tightly organised. If developers spend too much time on inefficient activities, the performance might decrease. The worst example could be breaks. Breaks could be seen as a reason for
economic problems by the management. This paper investigates the impact of breaks on the overall performance in software development. The investigations assume that if developers make pauses in a normal manner, this has not a negative impact on the profitability. In practice, the human-centric development process brings together a business process and a social process. The interaction of both processes was simulated. Due to the execution of 1.500 simulations, we obtained information on the economic progression of the development process under the influence of
breaks. We determined the impact of breaks on the overall profitability. This investigation contributes to the discussion of freedom in software development. It helps man-agers to
assess if employees who make breaks harm the profitability. This could lead to the imple-mentation of further business constraints.
Making decision in the presence of uncertainty requires estimating the impact of the outcome of those decisions. Here;s a collection of resources can can be used to guide that process
5.0 Estimating Agile Development ProjectsGlen Alleman
The primary purpose of software estimation is not to predict a project’s outcome; it is to determine whether a project’s targets are realistic enough to allow the project to be controlled to meet them ‒ Steve McConnell
SECURITY FOR DEVOPS DEPLOYMENT PROCESSES: DEFENSES, RISKS, RESEARCH DIRECTIONSijseajournal
DevOps is an emerging collection of software management practices intended to shorten time to market for
new software features and to reduce the risk of costly deployment errors. In this paper we examine the security implications of two of the key DevOps practices, automation of the deployment pipeline using a deployment toolchain and infrastructure-as-code to specify the environment of the eployed software. We focus on identifying what changes when an organization moves from manual deployments to DevOps
automated deployment processes.We reviewed the literature and conducted three case studies using simple configurations of common DevOps tools. This allowed us to identify specific:
• Positive influences on security where automation enhances defenses.
• Negative influences, where automation enables different kinds of attacks and increases the attack
surface.
• Research directions that look promising to support this new approach to software management.
• Recommendations for DevOps adopters
The primary purpose of software estimation is not to predict a project’s outcome; it is to determine whether a project’s targets are realistic enough to allow the project to be controlled to meet them ‒ Steve McConnell
Avoid software project horror stories - check the reality value of the estima...Harold van Heeringen
Many large software projects turn into software horror stories, resulting in newspaper headlines and even political issues. Often, the project costs and schedule were estimated unrealistically optimistic, using immature estimation techniques. A relatively simple way to avoid many problems is to perform a reality check on the estimate. This presentation was given on the conference of the International Cost Estimating and Analysis Association (ICEAA2014), June 2014 (Denver, USA)
Software Architecture Erosion: Impacts, Causes, and ManagementCSCJournals
As extensive new software projects are becoming more costly with increased development complexity and risks, the use of existing, already developed software projects are now spanning multiple years or decades. Software maintenance durations are increasing along with periodic periods of intense software upgrades due to these lengthy extensions of the life of the software. While the software architecture is set during the project's initial design phase, over the life of the project, architecture erosion can take place due to intentional and unintentional changes, resulting in deviations from the intended architecture. These deviations can manifest themselves in a wide variety of performance, maintenance, and software quality problems ranging from minor annoyances to product failure or unmaintainable software. This paper explores the causes and impacts of software erosion, and recovery approaches to addressing software erosion, laying the framework for future work towards the definition of an Architectural Maturity Model Integration process to assess the architectural maturity of an organization.
DevOps shifting software engineering strategy Value based perspectiveiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Preempting enterprise resource planning (ERP) project failure could well become the Holy Grail for organizations seeking to improve their technology infrastructure, organizational efficiency, and competitiveness. The business of preempting project failure begins with understanding that there is a preventative approach that can provide planned project assurance at critical points in the project’s evolution. It begins with clear understanding of expectations – from the executives, to the business and IT management, to the software vendors and end users. Learn how corporations are using project assurance methodologies to successfully implement enterprise software projects on time and on budget.
Unsustainable Regaining Control of Uncontrollable AppsCAST
The ever-growing cost to maintain systems continues to crush IT organizations robbing their ability to fund innovation while increasing risks across the organization. There are, however, some tactics to reduce application total ownership cost, reduce complexity and improve sustainability across your portfolio.
STRATEGIES TO REDUCE REWORK IN SOFTWARE DEVELOPMENT ON AN ORGANISATION IN MAU...ijseajournal
Rework is a known vicious circle in software development since it plays a central role in the generation of
delays, extra costs and diverse risks introduced after software delivery. It eventually triggers a negative
impact on the quality of the software developed. In order to cater the rework issue, this paper goes in depth
with the notion of rework in software development as it occurs in practice by analysing a development
process on an organisation in Mauritius where rework is a major issue. Meticulous strategies to reduce
rework are then analysed and discussed. The paper ultimately leads to the recommendation of the best
strategy that is software configuration management to reduce the rework problem in software development
The Increasing Value and Complexity of Software Call for the Reevaluation of ...PRQA
VDC performs annually global market research, regarding the Automated Test and Verification Tool (AT&V) market. The current white paper summarizes findings from a survey in 2011, capturing the input from almost 600 engineers. The respondents are directly involved in Embedded and Enterprise/IT software and systems
development, across a wide range of industries: Automotive, Medical, Telecoms, Aerospace, Military and others.
Programming Research (PRQA) worked with VDC to fine-tune the survey questions to provide deeper insights into the development of high integrity software and the role of static analysis.
Best practices for new project commissioningOlivia Wilson
Planning or implementing new projects is quite a complex task. If the commissioning process has been planned poorly or executed in a substandard fashion, it may make the project more complex and even result in unanticipated disruptions.
Best practices for new project commissioningOlivia Wilson
Planning or implementing new projects is quite a complex task. If the commissioning process has been planned poorly or executed in a substandard fashion, it may make the project more complex and even result in unanticipated disruptions.
SOFTWARE COST ESTIMATION USING FUZZY NUMBER AND PARTICLE SWARM OPTIMIZATIONIJCI JOURNAL
Software cost estimation is a process to calculate effort, time and cost of a project, and assist in better
decision making about the feasibility or viability of project. Accurate cost prediction is required to
effectively organize the project development tasks and to make economical and strategic planning, project
management. There are several known and unknown factors affect this process, so cost estimation is a very
difficult process. Software size is a very important factor that impacts the process of cost estimation.
Accuracy of cost estimation is directly proportional to the accuracy of the size estimation.
Failure of Software projects has always been an important area of focus for the Software Industry.
Implementation phase is not the only phase for Software projects to fail, instead planning and estimation
steps are the most crucial ones, which lead to their failure. More than 50% of the total projects fail which
go beyond the estimated time and cost. The Standish group‘s CHAOS reports failure rate of 70% for the
software projects. This paper presents the existing algorithms for software estimation and the relevant
concepts of Fuzzy Theory and PSO. Also explains the proposed algorithm with experimental results.
Accelerate Development by getting Requirements RightSerena Software
Requirements management is an often underutilized discipline in software development. Yet the economic benefits—lowered development costs, faster development and better products that delight customers—have plainly been demonstrated in study after study. In this paper, renowned engineering process expert Karl Wiegers makes the case for requirements management as an integral part of any product lifecycle. Metrics are provided to show the positive effect of requirements management, from individual developers to the business to the customer. Serena Dimensions RM and Serena Composer are examined as an example of how to bring effective requirements planning and management into product development.
STATE-OF-THE-ART IN EMPIRICAL VALIDATION OF SOFTWARE METRICS FOR FAULT PRONEN...IJCSES Journal
With the sharp rise in software dependability and failure cost, high quality has been in great demand.However, guaranteeing high quality in software systems which have grown in size and complexity coupled with the constraints imposed on their development has become increasingly difficult, time and resource consuming activity. Consequently, it becomes inevitable to deliver software that have no serious faults. In
this case, object-oriented (OO) products being the de facto standard of software development with their unique features could have some faults that are hard to find or pinpoint the impacts of changes. The earlier faults are identified, found and fixed, the lesser the costs and the higher the quality. To assess product quality, software metrics are used. Many OO metrics have been proposed and developed. Furthermore,
many empirical studies have validated metrics and class fault proneness (FP) relationship. The challenge is which metrics are related to class FP and what activities are performed. Therefore, this study bring together the state-of-the-art in fault prediction of FP that utilizes CK and size metrics. We conducted a systematic literature review over relevant published empirical validation articles. The results obtained are
analysed and presented. It indicates that 29 relevant empirical studies exist and measures such as complexity, coupling and size were found to be strongly related to FP.
Similar to How did Software Got So Reliable Without Proof? (20)
Cyaniclab : Software Development Agency Portfolio.pdfCyanic lab
CyanicLab, an offshore custom software development company based in Sweden,India, Finland, is your go-to partner for startup development and innovative web design solutions. Our expert team specializes in crafting cutting-edge software tailored to meet the unique needs of startups and established enterprises alike. From conceptualization to execution, we offer comprehensive services including web and mobile app development, UI/UX design, and ongoing software maintenance. Ready to elevate your business? Contact CyanicLab today and let us propel your vision to success with our top-notch IT solutions.
Software Engineering, Software Consulting, Tech Lead.
Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Security,
Spring Transaction, Spring MVC,
Log4j, REST/SOAP WEB-SERVICES.
Providing Globus Services to Users of JASMIN for Environmental Data AnalysisGlobus
JASMIN is the UK’s high-performance data analysis platform for environmental science, operated by STFC on behalf of the UK Natural Environment Research Council (NERC). In addition to its role in hosting the CEDA Archive (NERC’s long-term repository for climate, atmospheric science & Earth observation data in the UK), JASMIN provides a collaborative platform to a community of around 2,000 scientists in the UK and beyond, providing nearly 400 environmental science projects with working space, compute resources and tools to facilitate their work. High-performance data transfer into and out of JASMIN has always been a key feature, with many scientists bringing model outputs from supercomputers elsewhere in the UK, to analyse against observational or other model data in the CEDA Archive. A growing number of JASMIN users are now realising the benefits of using the Globus service to provide reliable and efficient data movement and other tasks in this and other contexts. Further use cases involve long-distance (intercontinental) transfers to and from JASMIN, and collecting results from a mobile atmospheric radar system, pushing data to JASMIN via a lightweight Globus deployment. We provide details of how Globus fits into our current infrastructure, our experience of the recent migration to GCSv5.4, and of our interest in developing use of the wider ecosystem of Globus services for the benefit of our user community.
Exploring Innovations in Data Repository Solutions - Insights from the U.S. G...Globus
The U.S. Geological Survey (USGS) has made substantial investments in meeting evolving scientific, technical, and policy driven demands on storing, managing, and delivering data. As these demands continue to grow in complexity and scale, the USGS must continue to explore innovative solutions to improve its management, curation, sharing, delivering, and preservation approaches for large-scale research data. Supporting these needs, the USGS has partnered with the University of Chicago-Globus to research and develop advanced repository components and workflows leveraging its current investment in Globus. The primary outcome of this partnership includes the development of a prototype enterprise repository, driven by USGS Data Release requirements, through exploration and implementation of the entire suite of the Globus platform offerings, including Globus Flow, Globus Auth, Globus Transfer, and Globus Search. This presentation will provide insights into this research partnership, introduce the unique requirements and challenges being addressed and provide relevant project progress.
Enterprise Resource Planning System includes various modules that reduce any business's workload. Additionally, it organizes the workflows, which drives towards enhancing productivity. Here are a detailed explanation of the ERP modules. Going through the points will help you understand how the software is changing the work dynamics.
To know more details here: https://blogs.nyggs.com/nyggs/enterprise-resource-planning-erp-system-modules/
A Comprehensive Look at Generative AI in Retail App Testing.pdfkalichargn70th171
Traditional software testing methods are being challenged in retail, where customer expectations and technological advancements continually shape the landscape. Enter generative AI—a transformative subset of artificial intelligence technologies poised to revolutionize software testing.
Innovating Inference - Remote Triggering of Large Language Models on HPC Clus...Globus
Large Language Models (LLMs) are currently the center of attention in the tech world, particularly for their potential to advance research. In this presentation, we'll explore a straightforward and effective method for quickly initiating inference runs on supercomputers using the vLLM tool with Globus Compute, specifically on the Polaris system at ALCF. We'll begin by briefly discussing the popularity and applications of LLMs in various fields. Following this, we will introduce the vLLM tool, and explain how it integrates with Globus Compute to efficiently manage LLM operations on Polaris. Attendees will learn the practical aspects of setting up and remotely triggering LLMs from local machines, focusing on ease of use and efficiency. This talk is ideal for researchers and practitioners looking to leverage the power of LLMs in their work, offering a clear guide to harnessing supercomputing resources for quick and effective LLM inference.
Unleash Unlimited Potential with One-Time Purchase
BoxLang is more than just a language; it's a community. By choosing a Visionary License, you're not just investing in your success, you're actively contributing to the ongoing development and support of BoxLang.
Quarkus Hidden and Forbidden ExtensionsMax Andersen
Quarkus has a vast extension ecosystem and is known for its subsonic and subatomic feature set. Some of these features are not as well known, and some extensions are less talked about, but that does not make them less interesting - quite the opposite.
Come join this talk to see some tips and tricks for using Quarkus and some of the lesser known features, extensions and development techniques.
In software engineering, the right architecture is essential for robust, scalable platforms. Wix has undergone a pivotal shift from event sourcing to a CRUD-based model for its microservices. This talk will chart the course of this pivotal journey.
Event sourcing, which records state changes as immutable events, provided robust auditing and "time travel" debugging for Wix Stores' microservices. Despite its benefits, the complexity it introduced in state management slowed development. Wix responded by adopting a simpler, unified CRUD model. This talk will explore the challenges of event sourcing and the advantages of Wix's new "CRUD on steroids" approach, which streamlines API integration and domain event management while preserving data integrity and system resilience.
Participants will gain valuable insights into Wix's strategies for ensuring atomicity in database updates and event production, as well as caching, materialization, and performance optimization techniques within a distributed system.
Join us to discover how Wix has mastered the art of balancing simplicity and extensibility, and learn how the re-adoption of the modest CRUD has turbocharged their development velocity, resilience, and scalability in a high-growth environment.
Custom Healthcare Software for Managing Chronic Conditions and Remote Patient...Mind IT Systems
Healthcare providers often struggle with the complexities of chronic conditions and remote patient monitoring, as each patient requires personalized care and ongoing monitoring. Off-the-shelf solutions may not meet these diverse needs, leading to inefficiencies and gaps in care. It’s here, custom healthcare software offers a tailored solution, ensuring improved care and effectiveness.
Field Employee Tracking System| MiTrack App| Best Employee Tracking Solution|...informapgpstrackings
Keep tabs on your field staff effortlessly with Informap Technology Centre LLC. Real-time tracking, task assignment, and smart features for efficient management. Request a live demo today!
For more details, visit us : https://informapuae.com/field-staff-tracking/
We describe the deployment and use of Globus Compute for remote computation. This content is aimed at researchers who wish to compute on remote resources using a unified programming interface, as well as system administrators who will deploy and operate Globus Compute services on their research computing infrastructure.
How Recreation Management Software Can Streamline Your Operations.pptxwottaspaceseo
Recreation management software streamlines operations by automating key tasks such as scheduling, registration, and payment processing, reducing manual workload and errors. It provides centralized management of facilities, classes, and events, ensuring efficient resource allocation and facility usage. The software offers user-friendly online portals for easy access to bookings and program information, enhancing customer experience. Real-time reporting and data analytics deliver insights into attendance and preferences, aiding in strategic decision-making. Additionally, effective communication tools keep participants and staff informed with timely updates. Overall, recreation management software enhances efficiency, improves service delivery, and boosts customer satisfaction.
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Accelerate Enterprise Software Engineering with PlatformlessWSO2
Key takeaways:
Challenges of building platforms and the benefits of platformless.
Key principles of platformless, including API-first, cloud-native middleware, platform engineering, and developer experience.
How Choreo enables the platformless experience.
How key concepts like application architecture, domain-driven design, zero trust, and cell-based architecture are inherently a part of Choreo.
Demo of an end-to-end app built and deployed on Choreo.
Accelerate Enterprise Software Engineering with Platformless
How did Software Got So Reliable Without Proof?
1. How Did Software Get So Reliable
Without Proof?
C.A.R. Hoare
Oxford University Computing Laboratory,
Wolfson Building, Parks Road, Oxford, OX1 3QD, UK
Abstract. By surveying current software engineering practice, this pa-
per reveals that the techniques employed to achieve reliability are little
different from those which have proved effectivein all other branches of
modern engineering: rigorous management of procedures for design in-
spection and review;quality assurance based on a wide range of targeted
tests; continuous evolution by removal of errors from products already in
widespread use; and defensive programming, among other forms of de-
liberate over-engineering. Formal methods and proof play a small direct
role in large scaleprogramming; but they do provide a conceptual frame-
work and basic understanding to promote the best of current practice,
and point directions for future improvement.
1 Introduction
Twenty years ago it was reasonable to predict that the size and ambition of soft-
ware products would be severely limited by the unreliability of their component
programs. Crude estimates suggest that professionally written programs deliv-
ered to the customer can contain between one and ten independently correctable
errors per thousand lines of code; and any software error in principle can have
spectacular effect (or worse: a subtly misleading effect) on the behaviour of the
entire system. Dire warnings have been issued of the dangers of safety-critical
software controlling health equipment, aircraft, weapons systems and industrial
processes, including nuclear power stations. The arguments were sufficientlyper-
suasive to trigger a significant research effort devoted to the problem of program
correctness. A proportion of this research was based on the ideal of certainty
achieved by mathematical proof.
Fortunately, the problem of program correctness has turned out to be far
less serious than predicted. A recent analysis by Mackenzie has shown that of
several thousand deaths so far reliably attributed to dependence on computers,
only ten or so can be explained by errors in the software: most of these were
due to a couple of instances of incorrect dosage calculations in the treatment
of cancer by radiation. Similarly predictions of collapse of software due to size
have been falsified by continuous operation of real-time software systems now
measured in tens of millions of lines of code, and subjected to thousands of
updates per year. This is the software which controls local and trunk telephone
exchanges; they have dramatically improved the reliability and performance of
2. telecommunications throughout the world. And aircraft, both civil and military,
are now flying with the aid of software measured in millions of lines - though
not all of it is safety-critical. Compilers and operating systems of a similar size
now number their satisfied customers in millions.
So the questions arise: why have twenty years of pessimistic predictions been
falsified? Was it due to successful application of the results of the research which
was motivated by the predictions? How could that be, when clearly little software
has ever has been subjected to the rigours of formal proof? The objective of these
enquiries is not to cast blame for the non-fulfilment of prophecies of doom. The
history of science and engineering is littered with false predictions and broken
promises; indeed they seem to serve as an essential spur to the advancement of
human knowledge; and nowadays, they are needed just to maintain a declining
flow of funds for research. Nixon's campaign to cure cancer within ten years
was a total failure; but it contributed in its time to the understanding on which
the whole of molecular medicine is now based. The proper role for an historical
enquiry is to draw lessons that may improve present practices, enhance the
accuracy of future predictions, and guide policies and directions for continued
research in the subject.
The conclusion of the enquiry will be that in spite of appearances, modern
software engineering practice owes a great deal to the theoretical concepts and
ideals of early research in the subject; and that techniques of formalisation and
proof have played an essential role in validating and progressing the research.
However, technology transfer is extremely slow in software, as it should be in any
serious branch of engineering. Because of the backlog of research results not yet
used, there is an immediate and continuing role for education, both of newcomers
to the profession and of experienced practitioners. The final recommendation is
that we must aim our future theoretical research on goals which are as far ahead
of the current state of the art as the current state of industrial practice lags
behind the research we did in the past. Twenty years perhaps?
2 Management
The most dramatic advances in the timely delivery of dependable software are
directly attributed to a wider recognition of the fact that the process of program
development can be predicted, planned, managed and controlled in the same
way as in any other branch of engineering. The eventual workings of the program
itself are internal to a computer and invisible to the naked eye; but that is no
longer any excuse for keeping the design process out of the view of management;
and the visibility should preferably extend to all management levels up to the
most senior. That is a necessary condition for the allocation of time, effort and
resources needed for the solution of longer term software problems like those of
reliability.
The most profitable investment of extra effort is known to be at the very start
of a project, beginning with an intensified study not only of the requirements
of the ultimate customer, but also of the relationship between the product and
3. the environment of its ultimate use. The greatest number (by far) of projects
that have ended in cancellation or failure in delivery and installation have already
begun to fail at this stage. Of course we have to live with the constant complaint
that the customers do not know what they want; and when at last they say they
do, they constantly change their mind. But that is no excuse for abrogating
management responsibility. Indeed, even stronger management is required to
explore and capture the true requirements, to set up procedures and deadlines
for management of change, to negotiate and where necessary invoke an early
cancellation clause in the contract. Above all, the strictest management is needed
to prevent premature commitment to start programming as soon as possible.
This can only lead to a volume of code of unknown and untestable utility, which
will act forever after as a dead weight, blighting the subsequent progress of the
project, if any.
The transition from an analysis of requirements to the specification of a pro-
gram to meet them is the most crucial stage in the whole project; the discovery
at this stage of only a single error or a single simplification would fully repay all
the effort expended. To ensure the proper direction of effort, the management
requires that all parts of the specification must he subjected to review by the
best and most experienced software architects, who thereby take upon them-
selves an appropriate degree of responsibility for the success of the project. That
is what enables large implementation teams to share the hard-won experience
and judgement of the best available engineers.
Such inspections, walkthroughs, reviews and gates are required to define im-
portant transitions between all subsequent phases in the project, from project
planning, design, code, test planning, and evaluation of test results. The individ-
ual designer or programmer has to accept the challenge not only of making the
right decisions, hut also of presenting to a group of colleagues the arguments and
reasons for confidence in their correctness. This is amazingly effective in instill-
ing and spreading a culture conducive to the highest reliability. Furthermore, if
the review committee is not satisfied that the project can safely proceed to its
next phase, the designer is required to re-work the design and present it again.
Even at the earliest stage, management knows immediately of the setback, and
already knows, even if they refuse to believe it, that the delivery will have to be
rescheduled by exactly the same interval that has been lost. Slack for one or two
such slippages should be built into the schedule; but if the slack is exhausted,
alternative and vigorous action should be no longer delayed.
At the present day, most of the discussion at review meetings is conducted
in an entirely informal way, using a language and conceptual framework evolved
locally for the purpose. However, there is now increasing experience of the ben-
efits of introducing abstract mathematical concepts and reasoning methods into
the process, right from the beginning. This permits the consequences of each
proposed feature and their possible combinations to be explored by careful and
exhaustive mathematical reasoning, to avoid the kind of awkward and perhaps
critical interactions that might otherwise be detected only on delivery. At the
design stage, the mathematics can help in exploring the whole of the design
4. space, and so give greater assurance that the simplest possible solution has been
adopted. Even stricter formalisation is recommended for specifying the interfaces
between the components of the design, to be implemented perhaps in different
places at different times by different people. Ideally, one would like to see a proof
in advance of the implementation that correctness of the components, defined
in terms of satisfaction of the interface specifications, will guarantee correctness
of their subsequent assembly. This can greatly reduce the risk of a lengthy and
unpredictable period of integration testing before delivery.
At the final review of the code, judicious use of commentary in the form
of assertions, preconditions, postconditions and invariants can greatly help in
marshalling a convincing argument that a program actually works. Furthermore,
it is much easier to find bugs in a line of reasoning than it is in a line of code. In
principle, correctness of each line of reasoning depends at most on two preceding
lines of reasoning, which are explicitly referenced. In principle, correctness of
each line of code depends on the behaviour of every other line of code in the
system.
Success in the use of mathematics for specification, design and code reviews
does not require strict formalisation of all the proofs. Informal reasoning among
those who are fluent in the idioms of mathematics is extremely efficient, and
remarkably reliable. It is not immune from failure; for example simple misprints
can be surprisingly hard to detect by eye. Fortunately, these are exactly the
kind of error that can be removed by early tests. More formal calculation can be
reserved for the most crucial issues, such as interrupts and recovery procedures,
where bugs would be most dangerous, expensive, and most difficult to diagnose
by tests.
A facility in formalisation and effective reasoning is only one of the talents
that can help in a successful review. There are many other less formal talents
which are essential. They include a wide understanding of the application area
and the marketplace, an intuitive sympathy with the culture and concerns of the
customer, a knowledge of the structure and style of existing legacy code, acquain-
tance and professional rapport with the most authoritative company experts on
each relevant topic, a sixth sense for the eventual operational consequences of
early design decisions, and above all, a deep sense of personal commitment to
quality, and the patience to survive long periods of intellectual drudgery needed
to achieve a thoroughly professional result. These attributes are essential. The
addition of mathematical fluency to the list is not going to be easy; the best
hope is to show that it will enhance performance in all these other ways as well.
3 Testing
Thorough testing is the touchstone of reliability in quality assurance and control
of modern production engineering. Tests are applied as early as possible at all
stations in the production line. They are designed rigorously to maximise the
likelihood of failure, and so detect a fault as soon as possible. For example, if
parameters of a production process vary continuously, they are tested at the
5. extreme of their intended operating range. Satisfaction of all tests in the fac-
tory affords considerably increased confidence, on the part of the designer, the
manufacturer, and the general public, that the product will continue to work
without fail throughout its service lifetime. And the confidence is justified: mod-
ern consumer durables are far more durable than they were only twenty years
ago.
But computing scientists and philosophers remain skeptical. E.W. Dijkstra
has pointed out that program testing can reveal only the presence of bugs, never
their absence. Philosophers of science have pointed out that no series of experi-
ments, however long and however favourable can ever prove a theory correct; but
even only a single contrary experiment will certainly falsify it. And it is a basic
slogan of quality assurance that "you cannot test quality into a product". How
then can testing contribute to reliability of programs, theories and products? Is
the confidence it gives illusory?
The resolution of the paradox is well known in the theory of quality control.
It is to ensure that a test made on a product is not a test of the product itself,
but rather of the methods that have been used to produce it - the processes,
the production lines, the machine tools, their parameter settings and operating
disciplines. If a test fails, it is not enough to mend the faulty product. It is not
enough just to throw it away, or even to reject the whole batch of products in
which a defective one is found. The first principle is that the whole production
line must be re-examined, inspected, adjusted or even closed until the root cause
of the defect has been found and eliminated.
Scientists are equally severe with themselves. To test a theory they devise
a series of the most rigorous possible experiments, aimed explicitly and exclu-
sively to disprove it. A single test with a negative result may occasionally be
attributed to impure ingredients or faulty apparatus; but if the negative out-
come is repeated, parts of the theory have to be rethought and recalculated;
when this gets too complicated, the whole theory has to be abandoned. As Pop-
per points out, the non-scientist will often die with (or even for) his false beliefs;
the scientist allows his beliefs to die instead of himself.
A testing strategy for computer programs must be based on lessons learned
from the successful treatment of failure in other branches of science and engi-
neering. The first lesson is that the test strategy must be laid out in advance
and in all possible detail at the very earliest stage in the planning of a project.
The deepest thought must be given to making the tests as severe as possible,
so that it is extremely unlikely that an error in the design of the program could
possibly remain undetected. Then, when the program is implemented and passes
all its tests the first time, it is almost unbelievable that there could be any in-
herent defect in the methods by which the program has been produced or any
systematic lapse in their application. This is the message of Harlan Mill's "clean
room" strategy.
The earliest possible design of the test strategy has several other advantages.
It encourages early exploration, simplification and clarification of the assump-
tions underlying use of the program, especially at edges of its operating range;
6. it facilitates early detection of ambiguities and awkward interaction effects la-
tent in the specification; and it concentrates attention from the earliest stage on
central problems of assuring correctness of the system as a whole. Many more
tests should be designed than there will ever be time to conduct; they should be
generated as directly as possible from the specification, preferably automatically
by computer program. Random selection at the last minute will protect against
the danger that under pressure of time the program will be adapted to pass the
tests rather than meeting the rest of its specification. There is some evidence
that early attention to a comprehensive and rigorous test strategy can improve
reliability of a delivered product, even when at the last minute there was no time
to conduct the tests before delivery!
The real value of tests is not that they detect bugs in the code, but that they
detect inadequacy in the methods, concentration and skills of those who design
and produce the code. Programmers who consistently fail to meet their test-
ing schedules are quickly isolated, and assigned to less intellectually demanding
tasks. The most reliable code is produced by teams of programmers who have
survived the rigours of testing and delivery to deadline over a period of ten years
or more. By experience, intuition, and a sense of personal responsibility they are
well qualified to continue to meet the highest standards of quality and reliability.
But don't stop the tests: they are still essential to counteract the distracting ef-
fects and the perpetual pressure of close deadlines, even on the most meticulous
programmers.
Tests that are planned before the code is written are necessarily "black box"
tests; they operate only at the outermost interfaces of the product as a whole,
without any cognizance of its internal structure. Black box tests also fulfil an
essential role as acceptance tests, for use on delivery of the product to the cus-
tomer's site. Since software is invisible, there is absolutely no other way of check-
ing that the version of the software loaded and initialised on the customer's ma-
chine is in fact the same as what has been ordered. Another kind of acceptance
test is the suite of certification tests which are required for implementations
of standard languages like COBOL and ADA. They do little to increase confi-
dence in the overall reliability of the compiler, but they do at least fairly well
ensure that all the claimed language features have in fact been delivered; past
experience shows that even this level of reliability cannot be taken for granted.
Another common kind of black box test is regression testing. When main-
taining a large system over a period of many years, all suggested changes are
submitted daily or weekly to a central site. They are all incorporated together,
and the whole system is recompiled, usually overnight or at the week end. But
before the system is used for further development, it is subjected to a large suite
of tests to ensure that it still works; if not, the previous version remains in use,
and the programmer who caused the error has an uncomfortable time until it is
mended. The regression tests include all those that have detected previous bugs,
particularly when this was done by the customer. Experience shows that bugs
are often a result of obscurity or complication in the code or its documentation;
and any new change to the code is all too likely to reintroduce the same bug -
something that customers find particularly irksome.
7. 4 Debugging
The secret of the success of testing is that it checks the quality of the process
and methods by which the code has been produced. These must be subjected to
continued improvement, until it is normal to expect that every test will be passed
first time, every time. Any residual lapse from this ideal must be tracked to
its source, and lead to lasting and widely propagated improvements in practice.
Expensive it may be, but that too is part of the cure. In all branches of commerce
and industry, history shows dramatic reduction in the error rates when their cost
is brought back from the customer to the perpetrator.
But there is an entirely different and very common response to the discov-
ery of an error by test: just correct the error and get on with the job. This is
known as debugging, by analogy with the attempt to get rid of an infestation
of mosquitoes by killing the ones that bite you - so much quicker and cheaper
and more satisfying than draining the swamps in which they breed. For insect
control, the swatting of individual bugs is known to be wholly ineffective. But
for programs it seems very successful; on removal of detected bugs, the rate of
discovery of new bugs goes down quite rapidly, at least to begin with. The reso-
lution of the paradox is quite simple; it is as if mosquitoes could be classified into
two very distinct populations, a gentle kind that hardly ever bite, and a vicious
kind that bite immediately. By encouraging the second kind, it is possible to
swat them, and then live comfortably with the yet more numerous swarm that
remains. It seems possible that a similar dichotomy in software bugs gives an
explanation of the effectiveness of debugging.
The first tests of newly written code are those conducted by the programmer
separately on isolated segments. These are extraordinarily effective in remov-
ing typographical errors, miskeying, and the results of misunderstanding the
complexity of the programming language, the run-time library or the operating
system. This is the kind of error that is easily made, even by the most com-
petent and diligent programmer, and fortunately just as easily corrected in to-
day's fast-turnround visual program debugging environments. Usually, the error
is glaringly obvious on the first occasion that a given line of code is executed.
For this reason, the objective of the initial test suite is to drive the program
to execute each line of its code at least once. This is known as a coverage test;
because it is constructed in complete knowledge of the object under test, it
is classified as an "open box" test. In hardware design a similar principle is
observed; the suite of tests must ensure that every stable element makes at least
one transition from high voltage to low and at least one transition from low
voltage to high. Then at least any element that is stuck at either voltage level
will be detected.
The cheapest way of testing a new or changed module of code in a large
system is simply to insert the module in the system and run the standard suite of
regression tests. Unfortunately, the coverage achieved in this way does not seem
8. adequate: the proportion of code executed in regression tests has been reported
to be less than thirty per cent. To improve this figure, a special test harness has
to be constructed to inject parameters and inspect results at the module level.
Unfortunately, for a module with many parameters, options and modes, to push
the coverage towards a hundred percent gets increasingly difficult; in the testing
of critical software for application in space, comprehensive testing is reported
to increase costs by four times as much as less rigorously tested code. Equally
unfortunately, total coverage is found to be necessary: more errors continue to
be discovered right up to the last line tested.
In hardware design, exhaustive testing of stuck-at faults has also become im-
possible, because no sufficientlysmall part of a chip can be exercised in isolation
from the rest. Nevertheless, quite short test sequences are adequate to identify
and discard faulty chips as they come off the production line. It is a fortunate
property of the technology of VLSI that any faults that are undetected by the
initial tests will very probably never occur; or at least they will never be noticed.
They play the role of the gentle kind of mosquito: however numerous, they hardly
ever bite.
Returning to the case of software, when the program or the programmer has
been exhausted by unit testing, the module is subjected to regression testing,
which may throw up another crop of errors. When these are corrected, the regres-
sion tests soon stop detecting new errors. The same happens when an updated
system is first delivered to the customer: nearly all the errors are thrown up
in early runs of the customer's own legacy code. After that, the rate at which
customers report new errors declines to a much lower and almost constant figure.
The reason for this is that even the most general-purpose programs are only
used in highly stereotyped ways, which exercise only a tiny proportion of the total
design space of possible paths through the code. Most of the actual patterns of
use are explored by the very first regression tests and legacy tests, and beta
testing enables the customer to help too. When the errors are removed from the
actually exercised paths, the rate at which new paths are opened up is very low.
Even when an anomaly is detected, it is often easier to avoid it by adapting the
code that invokes it; this can be less effort and much quicker than reporting the
error. Perhaps it is by this kind of mutual adaption that the components of a
large system, evolving over many years, reach a level of natural symbiosis; as
in the world of nature, the reliability and stability and robustness of the entire
system is actually higher than that of any of its parts.
When this stable state is reached, analysis of a typical error often leads to an
estimate that, even if the error were uncorrected, the circumstances in which it
occurs are so unlikely that on a statistical basis they will not occur again in the
next five thousand years. Suppose a hundred new errors of this kind are detected
each year. Crude extrapolation suggests that there must be about half a million
such errors in the code. Fortunately, they play the same role as the swarms of
the gentle kind of mosquito that hardly ever bites. The less fortunate corollary is
that if all the errors that are detected are immediately corrected, it would take
a thousand years to reduce the error rate by twenty percent. And that assumes
9. that there are no new errors introduced by the attempt to correct one which has
already been detected. After a certain stage, it certainly pays both the customer
and the supplier to leave such errors unreported and uncorrected.
Unfortunately, before that stage is reached, it often happens that a new
version of the system is delivered, and the error rate shoots up again. The costs
to the customer are accepted as the price of progress: the cost to the supplier is
covered by the profit on the price of the software. The real loss to the supplier is
the waste of the time and skill of the most experienced programmers, who would
otherwise be more profitably employed in implementing further improvements
in the functionality of the software. Although (surprisingly) the figures are often
not officially recorded, the programmers themselves estimate that nearly half
their time is spent in error correction. This is probably the strongest commercial
argument for software producers to increase investment in measures to control
reliability of delivered code.
5 Over-engineering
The concept of a safety factor is pervasive in engineering. After calculating the
worst case load on a beam, the civil engineer will try to build it ten times
stronger, or at least twice as strong, whenever the extra cost is affordable. In com-
puting, a continuing fall in price of computer storage and increase in computer
power has made almost any trade-off acceptable to reduce the risk of software
error, and the scale of damage that can increasingly result from it. This leads to
the same kind of over-engineering as is required by law for bridge-building; and
it is extremely effective, even though there is no clear way of measuring it by a
numeric factor.
The first benefit of a superabundance of resource is to make possible a de-
cision to avoid any kind of sophistication or optimisation in the design of algo-
rithms or data structures. Common prohibitions are: no data packing, no optimal
coding, no pointers, no sharing, no dynamic storage allocation. The maximum
conceivably necessary size of record or array is allocated, and then some more.
Similar prohibitions are often placed on program structures: no jumps, no inter-
rupts, no multiprogramming, no global variables. Access to data in other modules
is permitted only through carefully regulated remote procedure calls. In the past,
these design rules were found to involve excessive loss of efficiency; up to a factor
of a hundred has been recorded on first trials of a rigorously structured system.
This factor had to be regained by relaxing the prohibitions, massaging the inter-
faces between modules, even to the extent of violating the structural integrity
of the whole system. Apart from the obvious immediate dangers, this can lead
to even greater risk and expense in subsequent updating and enhancing of the
system. Fortunately, cheaper hardware reduces the concern for efficiency, and
improved optimisation technology for higher level languages promises further
assistance in reconciling a clear structure of the source code with high efficiency
in the object code.
Profligacy of resources can bring benefits in other ways. When considering
10. ]0
a possible exceptional case, the programmer may be quite confident that it has
already been discriminated and dealt with elsewhere in some other piece of code;
as a result in fact the exception can never arise at this point. Nevertheless, for
safety, it is better to discriminate again, and write further code to deal with
it. Most likely, the extra code will be totally unreachable. This may be part
of the explanation why in normal testing and operation, less than twenty per
cent of the code of a large system is ever executed; which suggests an over-
engineering factor of five. The extra cost in memory size may be low, but there
is a high cost in designing, writing and maintaining so much redundant code.
For example, there is the totally pointless exercise of designing coverage tests
for this otherwise unreachable code.
Another profligate use of resources is by cloning of code. A new feature to be
added to a large program can often be cheaply implemented by making a number
of small changes to some piece of code that is already there. But this is felt to
be risky: the existing code is perhaps used in ways that are not at all obvious
by just looking at it, and any of these ways might be disrupted by the proposed
change. So it seems safer to take an entirely fresh copy of the existing code,
and modify that instead. Over a period of years there arise a whole family of
such near-clones, extending over several generations. Each of them is a quick and
efficient solution to an immediate problem; but over time they create additional
problems of maintenance of the large volumes of code. For example, if a change
is made in one version of the clone, it is quite difficult even to decide whether
it should be propagated to the other versions, so it usually isn't. The expense
arises when the same error or deficiency has to be detected and corrected again
in the other versions.
Another widespread over-engineering practice is known as defensive program-
ming. Each individual programmer or team erects a defensive barrier against er-
rors and instabilities in the rest of the system. This may be nothing more than a
private library of subroutines through which all calls are made to the untrusted
features of a shared operating system. Or it may take the form of standard
coding practices. For example, it is recommended in a distributed system to
protect every communication with the environment, or with another program,
by a timeout, which will be invoked if the external response is not sufficiently
prompt. Conversely, every message accepted from the environment is subjected
to rigorous dynamic checks of plausibility, and the slightest suspicion will cause
the message to be just ignored, in the expectation that its sender is similarly
protected by timeout.
A similar technique can be applied to the global data structures used to
control the entire system. A number of checking programs, known as software
audits, are written to conduct plausibility checks on all the records in the global
system tables. In this case, suspicious entries are rendered harmless by a reini-
tialisation to safe values. Such audits have been found to improve mean time
between crashes of an embedded system from hours to months. The occasional
loss of data and function is unnoticed in a telephone switching application: it
could hardly be recommended for air traffic control, where it would certainly
11. 11
cause quite a different kind of crash.
The ultimate and very necessary defenee of a real time system against arbi-
trary hardware error or operator error is the organisation of a rapid procedure
for restarting the entire system. The goal of a restart is to restore the system
to a valid state that was current some time in the recent past. These warm
starts can be so efficient that they are hardly noticeable except by examining
the historical system log. So who cares whether the trigger for a restart was a
rare software fault or a transient hardware fault? Certainly, it would take far too
long to record information that would permit them to be discriminated.
The limitation of over-engineering as a safety technique is that the extra
weight and volume may begin to contribute to the very problem that it was
intended to solve. No-one knows how much of the volume of code of a large
system is due to over-engineering, or how much this costs in terms of reliability.
In general safety engineering, it is not unknown for catastrophes to be caused
by the very measures that are introduced to avoid them.
6 Programming Methodology
Most of the measures described so far for achieving reliability of programs are the
same as those which have proved to be equally effective in all engineering and in-
dustrial enterprises, from space travel to highway maintenance, from electronics
to the brewing of beer. But the best general techniques of management, quality
control, and safety engineering would be totally useless, by themselves; they are
only effective when there is a general understanding of the specific field of en-
deavour, and a common conceptual framework and terminology for discussion
of the relationship between cause and effect, between action and consequence in
that field. Perhaps initially, the understanding is based just on experience and
intuition; but the goal of engineering research is to complement and sometimes
replace these informal judgements by more systematic methods of calculation
and optimisation based on scientific theory.
Research into programming methodology has a similar goal, to establish a
conceptual framework and a theoretical basis to assist in systematic derivation
and justification of every design decision by a rational and explicable train of
reasoning. The primary method of research is to evaluate proposed reasoning
methods by their formalisation as a collection of proof rules in some completely
formal system. This permits definitive answers to the vital questions: is the
reasoning valid? is it adequate to prove everything that is needed? and is it
simpler than other equally valid and adequate alternatives? It is the provably
positive answer to these simple questions that gives the essential scientific basis
for a sound methodologicalreeornmendation- certainly an improvement on mere
rhetoric, speculation, fashion, salesmanship, charlatanism or worse.
Research into programming methodology has already had dramatic effects
on the way that people write programs today. One of the most spectacular
successes occurred so long ago that it is now quite non-controversial. It is the
almost universal adoption of the practice of structured programming, otherwise
12. 12
known as avoidance ofjumps (or gotos). Millions of lines of code have now been
written without them. But it was not always so. At one time, most programmers
were proud of their skill in the use ofjumps and labels. They regarded structured
notations as unnatural and counter-intuitive, and took it as a challenge to write
such complex networks ofjumps that no structured notations could ever express
them.
The decisive breakthrough in the adoption of structured programming by
IBM was the publication of a simple result in pure programming theory, the
Bohm-Jacopini theorem. This showed that an arbitrary program with jumps
could be executed by an interpreter written without any jumps at all; so in
principle any task whatsoever can be carried out by purely structured code.
This theorem was needed to convince senior managers of the company that no
harm would come from adopting structured programming as a company policy;
and project managers needed it to protect themselves from having to show their
programmers how to do it by rewriting every piece of complex spaghetti code
that might be submitted. Instead the programmers were just instructed to find
a way, secure in the knowledge that they always could. And after a while, they
always did.
The advantages of structured programming seem obvious to those who are
accustomed to it: programs become easy to write, to understand, and to modify.
But there is also a good scientific explanation for this judgement. It is found
by a formalisation of the methods needed to prove the correctness of the pro-
gram with the aid of assertions. For structured programs, a straightforward proof
always suffices. Jumps require a resort to a rather more complex technique of
subsidiary deductions. Formalisation has been invaluable in giving objective sup-
port for a subjective judgement: and that is a contribution which is independent
of any attempt to actually use the assertional proof rules in demonstrating the
correctness of code.
Another triumph of theory has been widespread appreciation of the benefits
of data types and strict type-checking of programs. A type defines the outer limits
of the range of values for a program variable or parameter. The range of facilities
for defining types is sufficiently restricted that a compiler can automatically
check that no variable strays outside the limits imposed by its declared type.
The repertoire of operations on the values of each type are defined by simple
axioms similar to those which define the relevant branch of mathematics. Strict
typechecking is certainly popular in Universities, because of the help it gives in
the teaching of programming to large classes of students with mixed abilities; it
is even more widely beneficial in modern mass consumer languages like Visual
Basic; and in very large programs which are subject to continuous change, it
gives a vital assurance of global system integrity that no programmer on the
project would wish to forego.
Another triumph of theoretical research has been widespread adoption of the
principles of information hiding. An early example is found in the local variables
of ALGOL 60. These are introduced by declaration and used as workspace for
internal purposes of a block of code which constitutes the scope of the declara-
13. 13
tion; the variable name, its identity, and even its existence is totally concealed
from outside. The concept of declaration and locality in a program was based
on that of quantification and bound variables in predicate logic; and so are the
proof methods for programs which contain them.
The information hiding introduced by the ALGOL 60 local variable was
generalised to the design of larger-scale modules and classes of object-oriented
programming, introduced into ALGOL 60 by SIMULA 67. Again, the scientific
basis of the structure was explored by formalisation of the relevant proof tech-
niques, involving an explicit invariant which links an abstract concept with its
concrete representation as data in the store of a computer.
The value of a foundation in formal logic and mathematics is illustrated by
the comparison of ALGOL 60 with the COBOL language, brought into existence
and standardised at about the same time by the U.S. Department of Defence.
Both languages had the highly commendable and explicit objective of making
programs easier to understand. COBOL tried to do this by constructing a crude
approximation to normal natural English, whereas ALGOL 60 tried to get closer
to the language of mathematics. There is no doubt which was technically more
successful: the ideas of ALGOL 60 have been adopted by many subsequent lan-
guages, including even FORTRAN 90. COBOL by comparison has turned out
to be an evolutionary dead end.
7 Conclusion
This review of programming methodology reveals how much the best of current
practice owes to the ideas and understanding gained by research which was
completed more than twenty years ago. The existence of such a large gap between
theory and practice is deplored by many, but I think quite wrongly. The gap is
actually an extremely good sign of the maturity and good health of our discipline,
and the only deplorable results are those that arise from failure to recognise it.
The proper response to the gap is to first congratulate the practitioners for
their good sense. Except in the narrowest areas, and for the shortest possible
periods of time, it would be crazy for industry to try to keep pace with the latest
results of pure research. If the research fails, the industry fails with it; and if
the research continues to succeed, the industry which is first to innovate runs
the risk of being overtaken by competitors who reap the benefits of the later
improvements. For these reasons, it would be grossly improper to recommend
industry on immediate implementation of results of their own research that is
still in progress. Indeed, Sir Richard Doll points out that scientists who give
such advice not only damage their clients; they also lose that most precious of
all attributes of good research, their scientific objectivity.
The theorists also should be accorded a full share of the congratulations;
for it is they who have achieved research results that are twenty years ahead of
the field of practice. It is not their failing but rather their duty to achieve and
maintain such an uncomfortable lead, and to spread it over a broad front across
a wide range of theories. No-one can predict, with any certainty or accuracy of
14. 14
detail, the timescales of change in technology or in the marketplace. The duty
of the researcher is not to predict the future more accurately than the busi-
nessman, but to prepare the basic understanding which may be needed to deal
with the unexpected challenges of any possible future development. Provided
that this goal has been met, no researcher should be blamed for failure of early
predictions made to justify its original funding of the research. Mistakes made
by businessmen and politicians are far more expensive.
The recognition of the appropriate timescale to measure the gap between the
theory and practice of a discipline is an essential to the appropriate planning of
research and education, both to fill the gap by improving practice, and to extend
it again by advancing the theory. I would recommend that the best researchers in
the field should simultaneously try to do both, because the influence of practice
on the development of theory is more beneficial and actually quicker than the
other way round.
At the extreme of the practical end, I would recommend the theorist to
alternate theoretical pursuits with much closer observation and experimentation
on actual working programs, with all the mass of documentation and historical
development logs that have accumulated in the last ten years. These systems
are now sufficiently stable, and have sufficient commercial prospects, to justify
quite practical research to answer questions that will guide recommendations for
future beneficial changes in their structure, content or methods of development.
For example, it would be very interesting to find a way of estimating the pro-
portional cost of cloning and the other over-engineering practices. By sampling,
it would be interesting to trace a number of errors to their root cause, and see
how they might have been avoided, perhaps by better specification or by better
documentation or by better structuring of code. Is my conjectured dichotomy
of error populations observed in practice? Any recommendation for improved
formalisation or improved structure will probably be based on other people's re-
search ideas that are up to twenty years old. Even so, they must be backed up by
trial recoding of a range of existing modules, selected on the scientific principle
of being the most likely to reveal the fallacies in the recommendation, rather
than its merits. Strange to relate, it has been known for a business to spend
many millions on a change that has not been subjected to any prior scientific
trials of this kind.
Formal methods researchers who are really keen on rigorous checking and
proof should identify and concentrate on the most critical areas of a large soft-
ware system, for example, synchronisation and mutual exclusion protocols, dy-
namic resource allocation, and reconfiguration strategies for recovery from par-
tial system failure. It is known that these are areas where obscure time-dependent
errors, deadlocks and livelocks (thrashing) can lurk untestable for many years,
and then trigger a failure costing many millions. It is possible that proof methods
and model checking are now sufficiently advanced that a good formal method-
ologist could occasionally detect such obscure latent errors before they occur in
practice. Publication of such an achievement would be a major milestone in the
acceptance of formal methods in solving the most critical problems of software
reliability.
15. 15
I have suggested that personal involvement in current practices and inspec-
tion of legacy code may lead to quite rapid benefits, both to the practitioner and
to the theorist. But this is not the right permanent relationship between them;
in a proper policy of technology transfer, it is for the practitioner to recognise
promising results of research, and take over all the hard work of adapting them
for widespread application. In software, unfortunately, the gap between practice
and theory is now so large that this is not happening. Part of the trouble is
that many or most of the practitioners did not study formal methods or even
computing science at University. This leaves a large educational gap, that can
only be filled by programme of in-service education which will acquaint some
of the best software engineers in industry with some of the important ideas of
computing science. Since many of them have degrees in mathematics, or at least
in some mathematical branch of science, they have the necessary background
and ability: since they do not have degrees in computing, they need to start
right at the beginning, for example, with context free languages and finite state
machines, and simple ideas of types and functional programming.
Another high barrier to technology transfer is the failure of software engi-
neering toolsets to include a modicum of support for formality - for example to
allow mathematical notations in word processors, to incorporate typechecking
for specifications, and hypertext techniques for quick cross-referencing between
formal and informal documentation. Improved tools should concentrate first on
very simple old techniques like execution profiles and selective compilation of
assertions before going on to more advanced but less mature technology, such as
model checking or proof assistance. The actual construction of industrial quality
tools must be done in collaboration with the industrial suppliers of these tools.
Only they have the knowledge and profit motive to adapt them, and to continue
adapting them, to the rapidly changing fashions and needs of the marketplace.
For long-term research, my advice is even more tentative and controversial.
It pursues a hope to complement the many strengths, and compensate the sin-
gle weakness, of current theoretical research in formal methods. The strengths
arise from the depth and the range of the specialisation of many flourishing re-
search schools in all the relevant areas. For example, in programming language
semantics, we have reasoning based on denotational, algebraic and operational
presentations. Among programming paradigms, we have both theoretical stud-
ies and applications of functional, procedural, logical and parallel programming
languages. Even among the parallel languages there is a great variation between
those based on synchronous or asynchronous control, shared store or distributed
message passing, untimed or with timing of various kinds; even hardware and
software have different models.
Specialisation involves a deep commitment to a narrow selection of presen-
tation, reasoning methods, paradigm, language and application area, or even a
particular application. The whole point of the specialisation in formal methods
is to restrict the notational framework as far as necessary to achieve some for-
mal goal, but nevertheless to show that the restrictions do not prevent successful
16. 16
application to a surprisingly wide range of problems. This is the reason why spe-
cialist research into formal methods can run the risk of being very divisive. An
individual researcher, or even a whole communityof researchers, becomes wholly
committed to a particular selection of specialisations along each of the axes: say
an operational or an algebraic presentation of semantics, bisimulation or term
rewriting as a proof method, CCS or OBJ as a design notation. The attraction of
such a choice can be well illustrated in certain applications, such as the analysis
of the alternating bit protocol or the definition of the stack as an abstract data
type. The perfectly proper challenge of the research is to push outwards as far as
possible the frontiers of the convenient application of the particular chosen for-
malism. But that is also the danger: the rush to colonise as much of the available
territory can lead to imperialist claims that deny to other specialisms their right
to existence. Any suggestion of variation of standard dogma is treated as akin
to treason. This tendency can be reinforced by the short-sightedness of funding
agencies~ which encourage exaggerated claims to the universal superiority of a
single notation and technique.
The consequences of the fragmentation of research into rival schools is in-
evitable: the theorists become more and more isolated, both from each other
and from the world of practice, where one thing is absolutely certain: that there
is no single cure for all diseases. There is no single theory for all stages of the
development of the software, or for all components even of a single application
program. Ideas, concepts, methods, and calculations will have to be drawn from
a wide range of theories, and they are going to have to work together consis-
tently, with no risk of misunderstanding, inconsistency or error creeping in at the
interfaces. One effective way to break formal barriers is for the best theorists to
migrate regularly between the research schools, in the hope that results obtained
in one research specialisation can be made useful in a manner acceptable by the
other. The interworking of theories and paradigms can also be explored from the
practical end by means of the case study, chosen as a simplified version of some
typical application. In my view, a case study that constructs a link between two
or more theories, used for different purposes at different levels of abstraction,
will be more valuable than one which merely presents a single formalisation, in
the hope that its merits, compared with rival formalisations, will be obvious.
They usually are, but unfortunately only to the author.
Since theories will have to be unified in application, the best help that ad-
vanced research can give is to unify them in theory first. Fortunately, unification
is something that theoretical research is very good at, and the way has been
shown again and again in both science and mathematics. Examples from science
include the discovery of the atomic theory of matter as a unified framework for
all the varied elements and components of chemistry; similarly, the gravitational
field assimilates the movement of the planets in the sky and cannon balls on
earth. In mathematics, we see how topology unifies the study of continuity in
all the forms encountered in geometry and analysis, how logic explains the valid
methods of reasoning in all branches of mathematics. I would suggest the current
strength of individual specialisation in theoretical computing science should be
17. 17
balanced by a commitment from the best and most experienced researchers to
provide a framework in which all the specialisations can be seen as just aspects
or variations of the same basic ideas. Then it will be clear how both existing
and new specialisations are all equally worthy of effort to deepen the theory or
broaden its application. But the aim is no longer to expand and colonise the
whole space but rather to find the natural boundaries at which one theory can
comfortably coexist and cooperate with its neighbours. Closing a gap between
one theory and another is just as important as closing the gap between theory
and practice; and just as challenging.
8 Acknowledgments
I am very grateful to many programmers and managers working in industry, who
have made available to me the benefits of their judgment and long experience.
In particular, I would like to praise the leading practitioners of the state of
the art in IBM at Hursley, in BNK at Maidenhead, and in Digital at Nashua.
Many contributions to my thinking are due to members of IFIP WG2.3 on
Programming Methodology, and to its chairman Cliff Jones, who made useful
suggestions to the previous draft of the paper. Finally, thanks to those named
in the paper, with apologies for lack of more formal reference.