1) The document discusses the application of S-shaped logistic growth curves to model and forecast technological trends over time.
2) It specifically fits a logistic curve to data on annual TRIZ publications from 1996-2006 to illustrate the three parameters (κ, α, β) of the simple logistic model.
3) The ceiling parameter κ represents the expected maximum number of future publications, the growth period parameters α and β specify the linear-like growth phase, and tm indicates the midpoint time of the symmetric S-curve.
Characteristics of Fuzzy Wheel Graph and Hamilton Graph with Fuzzy Ruleijtsrd
Graph theory is the concepts used to study and model various application in different areas. We proposed the wheel graph with n vertices can be defined as 1 skeleton of on n 1 gonal pyramid it is denoted by w nwith n 1 vertex n=3 . A wheel graph is hamiltonion, self dual and planar. In the mathematical field of graph theory, and a Hamilton path or traceable graph is a path in an undirected or directed graph that visits each vertex exactly once. A Hamiltonian cycle is a hamiltonian path that is a cycle. In this paper, we consider the wheel graph and also the hamilton graph using if then rules fuzzy numbers. The results are related to the find the degree of odd vertices and even vertices are same by applying if then rules through the paths described by fuzzy numbers. Nisha. D | Srividhya. B "Characteristics of Fuzzy Wheel Graph and Hamilton Graph with Fuzzy Rule" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-6 , October 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29319.pdf Paper URL: https://www.ijtsrd.com/mathemetics/other/29319/characteristics-of-fuzzy-wheel-graph-and-hamilton-graph-with-fuzzy-rule/nisha-d
Detecting and Reporting Extensional Concept Drift in Statistical Linked DataAlbert Meroño-Peñuela
This document discusses detecting and reporting extensional concept drift in statistical linked data. It presents a method to detect when the individuals belonging to two variants of the same concept have significantly different populations, indicating extensional drift. It also recommends annotating such drifts using SPARQL UPDATE queries to push results back to endpoints. As a case study, the method is applied to detect drift between occupations in Dutch historical censuses from 1889 and 1899, finding drift in 19.35% of concepts.
This document discusses using a multiplicative time series model (p,d,q)x(P,D,Q)s to forecast air transport demand.
It analyzes passenger data from Yemenia Airlines in Yemen over many years. The data shows seasonality with a 12-month period. To account for this, the model includes polynomials in the backward shift operator B with period s=12.
The model is fitted to the logarithms of the monthly passenger totals, which minimizes residuals. This confirms the appropriateness of a logarithmic transformation for this time series data. The model provides a means of linking observations within and across years to generate forecasts of air transport demand.
Coordinates are an ordered pair of numbers written in brackets separated by a comma that pinpoint the location of a point or shape on a graph by showing the number of places to move across the horizontal axis (the x coordinate) and up the vertical axis (the y coordinate), with the order they are written being important.
The document provides an introduction to integration. It explains that integration is the reverse of differentiation and can be used to find areas, volumes, and other quantities. It discusses how integration works by dividing a curve into infinitesimally thin slices and adding them up. The key concepts of the integral sign, the integrand, and the constant of integration are introduced. Applications of integration in physics, chemistry, medicine, and other fields are briefly outlined.
In this video Data Graphics has been discussed. How the data can be presented with the help of different line graph, poly graph, bar diagram, histogram and Scatter plot and semi logarithmic plot/graph.
Portion completed:
1.DATA GRAPHICS
2. REPRESENTATION OF DATA
3. line graph,
4. poly graph,
5. bar diagram,
6. histogram
7. Pie diagram
8. Wind rose and star diagram
9. Flow Charts
10. Simple Bar Diagram
11. Line and Bar Graph
12. Multiple Bar Diagram
13. Compound Bar Diagram
14. Pie Diagram
15. Scatter plot
16. Semi-log plot
Portland oregon riders monthly data Using RMeghna Baid
This document analyzes a time series dataset of monthly rider counts for Portland's public transportation system from 1960-1969. It describes:
- Contributions from various authors for coding, analysis, modeling, and presentation.
- The methodology used, including exploratory data analysis, time series decomposition to analyze trends and seasonality, checking for stationarity, ARIMA modeling, and accuracy validation through forecasting.
- Key results, such as the time series having a trend and seasonality, the residuals being stationary, an optimal ARIMA(1,0,1) model being identified with good accuracy based on MAPE and validation metrics.
This document outlines an introduction to R graphics using ggplot2 presented by the Harvard MIT Data Center. The presentation introduces key concepts in ggplot2 including geometric objects, aesthetic mappings, statistical transformations, scales, faceting, and themes. It uses examples from the built-in mtcars dataset to demonstrate how to create common plot types like scatter plots, box plots, and regression lines. The goal is for students to be able to recreate a sample graphic by the end of the workshop.
Characteristics of Fuzzy Wheel Graph and Hamilton Graph with Fuzzy Ruleijtsrd
Graph theory is the concepts used to study and model various application in different areas. We proposed the wheel graph with n vertices can be defined as 1 skeleton of on n 1 gonal pyramid it is denoted by w nwith n 1 vertex n=3 . A wheel graph is hamiltonion, self dual and planar. In the mathematical field of graph theory, and a Hamilton path or traceable graph is a path in an undirected or directed graph that visits each vertex exactly once. A Hamiltonian cycle is a hamiltonian path that is a cycle. In this paper, we consider the wheel graph and also the hamilton graph using if then rules fuzzy numbers. The results are related to the find the degree of odd vertices and even vertices are same by applying if then rules through the paths described by fuzzy numbers. Nisha. D | Srividhya. B "Characteristics of Fuzzy Wheel Graph and Hamilton Graph with Fuzzy Rule" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-6 , October 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29319.pdf Paper URL: https://www.ijtsrd.com/mathemetics/other/29319/characteristics-of-fuzzy-wheel-graph-and-hamilton-graph-with-fuzzy-rule/nisha-d
Detecting and Reporting Extensional Concept Drift in Statistical Linked DataAlbert Meroño-Peñuela
This document discusses detecting and reporting extensional concept drift in statistical linked data. It presents a method to detect when the individuals belonging to two variants of the same concept have significantly different populations, indicating extensional drift. It also recommends annotating such drifts using SPARQL UPDATE queries to push results back to endpoints. As a case study, the method is applied to detect drift between occupations in Dutch historical censuses from 1889 and 1899, finding drift in 19.35% of concepts.
This document discusses using a multiplicative time series model (p,d,q)x(P,D,Q)s to forecast air transport demand.
It analyzes passenger data from Yemenia Airlines in Yemen over many years. The data shows seasonality with a 12-month period. To account for this, the model includes polynomials in the backward shift operator B with period s=12.
The model is fitted to the logarithms of the monthly passenger totals, which minimizes residuals. This confirms the appropriateness of a logarithmic transformation for this time series data. The model provides a means of linking observations within and across years to generate forecasts of air transport demand.
Coordinates are an ordered pair of numbers written in brackets separated by a comma that pinpoint the location of a point or shape on a graph by showing the number of places to move across the horizontal axis (the x coordinate) and up the vertical axis (the y coordinate), with the order they are written being important.
The document provides an introduction to integration. It explains that integration is the reverse of differentiation and can be used to find areas, volumes, and other quantities. It discusses how integration works by dividing a curve into infinitesimally thin slices and adding them up. The key concepts of the integral sign, the integrand, and the constant of integration are introduced. Applications of integration in physics, chemistry, medicine, and other fields are briefly outlined.
In this video Data Graphics has been discussed. How the data can be presented with the help of different line graph, poly graph, bar diagram, histogram and Scatter plot and semi logarithmic plot/graph.
Portion completed:
1.DATA GRAPHICS
2. REPRESENTATION OF DATA
3. line graph,
4. poly graph,
5. bar diagram,
6. histogram
7. Pie diagram
8. Wind rose and star diagram
9. Flow Charts
10. Simple Bar Diagram
11. Line and Bar Graph
12. Multiple Bar Diagram
13. Compound Bar Diagram
14. Pie Diagram
15. Scatter plot
16. Semi-log plot
Portland oregon riders monthly data Using RMeghna Baid
This document analyzes a time series dataset of monthly rider counts for Portland's public transportation system from 1960-1969. It describes:
- Contributions from various authors for coding, analysis, modeling, and presentation.
- The methodology used, including exploratory data analysis, time series decomposition to analyze trends and seasonality, checking for stationarity, ARIMA modeling, and accuracy validation through forecasting.
- Key results, such as the time series having a trend and seasonality, the residuals being stationary, an optimal ARIMA(1,0,1) model being identified with good accuracy based on MAPE and validation metrics.
This document outlines an introduction to R graphics using ggplot2 presented by the Harvard MIT Data Center. The presentation introduces key concepts in ggplot2 including geometric objects, aesthetic mappings, statistical transformations, scales, faceting, and themes. It uses examples from the built-in mtcars dataset to demonstrate how to create common plot types like scatter plots, box plots, and regression lines. The goal is for students to be able to recreate a sample graphic by the end of the workshop.
This document describes proposed changes to improve a model for estimating project S-curves based on project attributes and conditions. The key changes are:
1. Changing the model outputs from the polynomial function's parameters (a, b) to the inflection point position (p) and slope (s), which better indicate schedule performance.
2. Adding two new input factors - project difficulty and participant competence - to capture schedule influences beyond basic attributes like cost and duration.
3. Using fuzzy inference systems instead of neural networks to build transparent input-output relationships from historical project data, applying fuzzy clustering and hybrid training.
The goal is to develop a model that more accurately predicts project progress curves by incorporating schedule performance indicators
The document discusses an S-curve model that relates per-capita income to insurance penetration. It finds:
1) Estimating life and non-life insurance penetration globally yields an S-curve, where income elasticity starts and ends at 1 but exceeds 1 at intermediate income levels.
2) For life insurance, the income at maximum elasticity is $15,000, while for non-life it is $10,000.
3) Using purchasing power parities rather than market exchange rates increases estimated penetration levels and elasticities for developing countries.
This document discusses technology S-curves and their implications for renewable energy alternatives. It finds that both wind and geothermal energy are poised to become more economical than fossil fuels relatively soon based on historical data. Government R&D funding has been lower for wind and geothermal compared to solar, and funding for fossil fuels may be excessive given their diminishing performance. Analyzing renewable technologies through an S-curve lens provides insights not found using other approaches and has implications for future government and industry investment decisions.
This document discusses three frameworks used by Gartner Group to analyze information systems research: the technology maturity curve, adoption curve, and identification of strategic technologies. The maturity curve tracks how a technology matures over time through various stages from embryonic to obsolescence. The adoption curve shows how technologies are adopted cumulatively by organizations over time. Considering where technologies fall on these curves can provide insights into appropriate research questions and methodologies. Identifying strategic technologies may help determine promising areas for new research.
The document provides documentation and a tutorial for implementing an S-curve motion profile in a MyoStat Motion Control system. An S-curve allows for smooth acceleration and deceleration during motion to prevent damage. The K69 parameter controls the S-curve function, with higher values creating a more pronounced curve. The tutorial instructs the user to set parameters, collect speed data, and graph results at different K69 values to observe the S-curve motion profile.
This document describes proposed changes to improve a model for estimating project S-curves based on project attributes and conditions. The key changes are:
1. Changing the model outputs from the polynomial function's parameters (a, b) to the inflection point position (p) and slope (s), which better indicate schedule performance.
2. Adding two new input factors - project difficulty and participant competence - to capture schedule influences beyond basic attributes like cost and duration.
3. Using fuzzy inference systems instead of neural networks to build transparent input-output relationships from historical project data, applying fuzzy clustering and hybrid training.
The goal is to develop a model that more accurately predicts project progress curves by incorporating schedule performance indicators
This document discusses the application of S-shaped curves, also known as logistic curves or S-curves, to model the evolution of systems over time. It provides background on the origin and development of S-curves as models of growth. S-curves have been widely used across many domains to describe trends like population growth, market penetration of new technologies, and diffusion of innovations. The document reviews several examples of S-curve applications and discusses their use in areas like technological forecasting and TRIZ problem solving. It argues that S-curves provide forecasting power because growth is ultimately limited by scarce resources based on mathematical concepts like Verhulst's logistic growth equation.
Tinjauan pustaka membahas konsep-konsep dasar listrik seperti arus listrik, tegangan, hambatan, hukum Ohm, dan pengukuran parameter listrik menggunakan alat pengukur seperti voltmeter dan amperemeter. Dibahas pula jenis-jenis resistor dan susunan resistor dalam rangkaian listrik.
The document provides documentation and a tutorial for implementing an S-curve motion profile in a MyoStat Motion Control system. An S-curve allows for smooth acceleration and deceleration during motion to prevent damage. The K69 parameter controls the S-curve function, with higher values creating a more pronounced curve. The tutorial instructs the user to set parameters, collect speed data, and graph results at different K69 values to observe the S-curve motion profile.
Ringkasan dokumen tersebut adalah:
1) Dokumen tersebut membahas tentang konsep dan teori kemiskinan menurut pandangan Islam dan model-model kemiskinan.
2) Pendidikan dianggap sebagai kebutuhan dasar manusia menurut model kebutuhan Galtung.
3) Terdapat berbagai faktor yang dapat menyebabkan terperangkapnya seseorang dalam kemiskinan seperti rendahnya pendidikan dan kesehatan.
The document discusses S curves, which are graphs of cumulative quantities like man-hours or costs plotted against time. It describes different types of S curves including baseline, target, and actual S curves. Baseline S curves show planned progress, target S curves show expected progress if tasks are completed as scheduled, and actual S curves show progress to date based on percentage completion. The document also discusses how to interpret and generate different types of S curves to analyze project schedule performance, progress, slippage, and other metrics.
It is a recent development in the field of Computer science, used to encode information within an abstract picture.Even though it provides same level of security as the bar codes, it ensures encoding of more amount of data as compared to the traditional ways of encoding.
This document discusses the application of S-shaped curves, also known as logistic curves or S-curves, to model the evolution of systems over time. It provides background on the origin and development of S-curves as models of growth. S-curves have been widely used across many domains to describe trends like population growth, market penetration of new technologies, and diffusion of innovations. The document reviews several examples of applying S-curves and discusses their use in fields like economics, biology, technology forecasting, and TRIZ. It argues that S-curves can provide forecasting power because growth is ultimately limited by scarce resources based on concepts like Malthusian growth models.
Analyzing The Term Structure Of Interest Rates Using The Dynamic Nelson Siege...Dawn Cook
This document summarizes a discussion paper that introduces time-varying parameters into the dynamic Nelson-Siegel model for analyzing the term structure of interest rates. Specifically, it allows the factor loadings in the Nelson-Siegel model to vary over time using step functions and spline functions, and introduces time-varying volatility specified as a spline function of time. Estimating the model on U.S. Treasury yield data, it finds the time-varying specifications significantly improve model fit compared to the standard dynamic Nelson-Siegel model which assumes constant loadings and volatility. It also illustrates how the model can handle an unbalanced dataset with missing observations.
Options on Quantum Money: Quantum Path- Integral With Serial ShocksAM Publications,India
The author previously developed a numerical multivariate path-integral algorithm, PATHINT, which has been applied to several classical physics systems, including statistical mechanics of neocortical interactions, options in financial markets, and other nonlinear systems including chaotic systems. A new quantum version, qPATHINT, has the ability to take into account nonlinear and time-dependent modifications of an evolving system. qPATHINT is shown to be useful to study some aspects of serial changes to systems. Applications to options on quantum money and blockchains in financial markets are discussed.
The document discusses patterns of technological substitution that challenge the traditional S-curve model. It presents several historical examples that demonstrate more complex substitution patterns than the smooth S-curve, including: concatenated generations in steelmaking technologies; overlapping generations in IBM mainframe computers; and a case of long-term feedback reversing the substitution of DDT as an insecticide due to environmental concerns. The author argues that accounting for these complex real-world patterns requires broadening the theoretical framework for understanding technological substitutions.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology
This document describes a Kriging component for spatial interpolation of climatological variables in the OMS modeling framework. Kriging is a geostatistical technique that interpolates values based on measured data and the spatial autocorrelation between data points. The component implements ordinary and detrended Kriging algorithms using 10 semivariogram models. It can interpolate both raster and point data and outputs the interpolated climatological variable values. Links are provided for downloading the component code, data, and OMS project files needed to run the interpolation.
This document describes proposed changes to improve a model for estimating project S-curves based on project attributes and conditions. The key changes are:
1. Changing the model outputs from the polynomial function's parameters (a, b) to the inflection point position (p) and slope (s), which better indicate schedule performance.
2. Adding two new input factors - project difficulty and participant competence - to capture schedule influences beyond basic attributes like cost and duration.
3. Using fuzzy inference systems instead of neural networks to build transparent input-output relationships from historical project data, applying fuzzy clustering and hybrid training.
The goal is to develop a model that more accurately predicts project progress curves by incorporating schedule performance indicators
The document discusses an S-curve model that relates per-capita income to insurance penetration. It finds:
1) Estimating life and non-life insurance penetration globally yields an S-curve, where income elasticity starts and ends at 1 but exceeds 1 at intermediate income levels.
2) For life insurance, the income at maximum elasticity is $15,000, while for non-life it is $10,000.
3) Using purchasing power parities rather than market exchange rates increases estimated penetration levels and elasticities for developing countries.
This document discusses technology S-curves and their implications for renewable energy alternatives. It finds that both wind and geothermal energy are poised to become more economical than fossil fuels relatively soon based on historical data. Government R&D funding has been lower for wind and geothermal compared to solar, and funding for fossil fuels may be excessive given their diminishing performance. Analyzing renewable technologies through an S-curve lens provides insights not found using other approaches and has implications for future government and industry investment decisions.
This document discusses three frameworks used by Gartner Group to analyze information systems research: the technology maturity curve, adoption curve, and identification of strategic technologies. The maturity curve tracks how a technology matures over time through various stages from embryonic to obsolescence. The adoption curve shows how technologies are adopted cumulatively by organizations over time. Considering where technologies fall on these curves can provide insights into appropriate research questions and methodologies. Identifying strategic technologies may help determine promising areas for new research.
The document provides documentation and a tutorial for implementing an S-curve motion profile in a MyoStat Motion Control system. An S-curve allows for smooth acceleration and deceleration during motion to prevent damage. The K69 parameter controls the S-curve function, with higher values creating a more pronounced curve. The tutorial instructs the user to set parameters, collect speed data, and graph results at different K69 values to observe the S-curve motion profile.
This document describes proposed changes to improve a model for estimating project S-curves based on project attributes and conditions. The key changes are:
1. Changing the model outputs from the polynomial function's parameters (a, b) to the inflection point position (p) and slope (s), which better indicate schedule performance.
2. Adding two new input factors - project difficulty and participant competence - to capture schedule influences beyond basic attributes like cost and duration.
3. Using fuzzy inference systems instead of neural networks to build transparent input-output relationships from historical project data, applying fuzzy clustering and hybrid training.
The goal is to develop a model that more accurately predicts project progress curves by incorporating schedule performance indicators
This document discusses the application of S-shaped curves, also known as logistic curves or S-curves, to model the evolution of systems over time. It provides background on the origin and development of S-curves as models of growth. S-curves have been widely used across many domains to describe trends like population growth, market penetration of new technologies, and diffusion of innovations. The document reviews several examples of S-curve applications and discusses their use in areas like technological forecasting and TRIZ problem solving. It argues that S-curves provide forecasting power because growth is ultimately limited by scarce resources based on mathematical concepts like Verhulst's logistic growth equation.
Tinjauan pustaka membahas konsep-konsep dasar listrik seperti arus listrik, tegangan, hambatan, hukum Ohm, dan pengukuran parameter listrik menggunakan alat pengukur seperti voltmeter dan amperemeter. Dibahas pula jenis-jenis resistor dan susunan resistor dalam rangkaian listrik.
The document provides documentation and a tutorial for implementing an S-curve motion profile in a MyoStat Motion Control system. An S-curve allows for smooth acceleration and deceleration during motion to prevent damage. The K69 parameter controls the S-curve function, with higher values creating a more pronounced curve. The tutorial instructs the user to set parameters, collect speed data, and graph results at different K69 values to observe the S-curve motion profile.
Ringkasan dokumen tersebut adalah:
1) Dokumen tersebut membahas tentang konsep dan teori kemiskinan menurut pandangan Islam dan model-model kemiskinan.
2) Pendidikan dianggap sebagai kebutuhan dasar manusia menurut model kebutuhan Galtung.
3) Terdapat berbagai faktor yang dapat menyebabkan terperangkapnya seseorang dalam kemiskinan seperti rendahnya pendidikan dan kesehatan.
The document discusses S curves, which are graphs of cumulative quantities like man-hours or costs plotted against time. It describes different types of S curves including baseline, target, and actual S curves. Baseline S curves show planned progress, target S curves show expected progress if tasks are completed as scheduled, and actual S curves show progress to date based on percentage completion. The document also discusses how to interpret and generate different types of S curves to analyze project schedule performance, progress, slippage, and other metrics.
It is a recent development in the field of Computer science, used to encode information within an abstract picture.Even though it provides same level of security as the bar codes, it ensures encoding of more amount of data as compared to the traditional ways of encoding.
This document discusses the application of S-shaped curves, also known as logistic curves or S-curves, to model the evolution of systems over time. It provides background on the origin and development of S-curves as models of growth. S-curves have been widely used across many domains to describe trends like population growth, market penetration of new technologies, and diffusion of innovations. The document reviews several examples of applying S-curves and discusses their use in fields like economics, biology, technology forecasting, and TRIZ. It argues that S-curves can provide forecasting power because growth is ultimately limited by scarce resources based on concepts like Malthusian growth models.
Analyzing The Term Structure Of Interest Rates Using The Dynamic Nelson Siege...Dawn Cook
This document summarizes a discussion paper that introduces time-varying parameters into the dynamic Nelson-Siegel model for analyzing the term structure of interest rates. Specifically, it allows the factor loadings in the Nelson-Siegel model to vary over time using step functions and spline functions, and introduces time-varying volatility specified as a spline function of time. Estimating the model on U.S. Treasury yield data, it finds the time-varying specifications significantly improve model fit compared to the standard dynamic Nelson-Siegel model which assumes constant loadings and volatility. It also illustrates how the model can handle an unbalanced dataset with missing observations.
Options on Quantum Money: Quantum Path- Integral With Serial ShocksAM Publications,India
The author previously developed a numerical multivariate path-integral algorithm, PATHINT, which has been applied to several classical physics systems, including statistical mechanics of neocortical interactions, options in financial markets, and other nonlinear systems including chaotic systems. A new quantum version, qPATHINT, has the ability to take into account nonlinear and time-dependent modifications of an evolving system. qPATHINT is shown to be useful to study some aspects of serial changes to systems. Applications to options on quantum money and blockchains in financial markets are discussed.
The document discusses patterns of technological substitution that challenge the traditional S-curve model. It presents several historical examples that demonstrate more complex substitution patterns than the smooth S-curve, including: concatenated generations in steelmaking technologies; overlapping generations in IBM mainframe computers; and a case of long-term feedback reversing the substitution of DDT as an insecticide due to environmental concerns. The author argues that accounting for these complex real-world patterns requires broadening the theoretical framework for understanding technological substitutions.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology
This document describes a Kriging component for spatial interpolation of climatological variables in the OMS modeling framework. Kriging is a geostatistical technique that interpolates values based on measured data and the spatial autocorrelation between data points. The component implements ordinary and detrended Kriging algorithms using 10 semivariogram models. It can interpolate both raster and point data and outputs the interpolated climatological variable values. Links are provided for downloading the component code, data, and OMS project files needed to run the interpolation.
A Review Of Major Paradigms And Models For The Design Of Civil Engineering Sy...Tracy Morgan
The document discusses the evolution of paradigms for designing civil engineering systems. It identifies five classical paradigms (A through E) and proposes a new emerging paradigm (F):
1. The empirical model (Paradigm A), static descriptive deterministic method (Paradigm B), static stochastic descriptive model (Paradigm C), and dynamic stochastic descriptive model (Paradigm D) describe previous approaches.
2. The new proposed paradigm is the interactive multi-attribute learning paradigm (Paradigm F), which involves identifying actors, modeling structures, and using tools to generate alternatives, simulate responses, communicate features, and assess reactions and preferences in an interactive learning process.
Simulating Multivariate Random Normal Data using Statistical Computing Platfo...ijtsrd
This document describes how to simulate multivariate normal random data using the statistical computing platform R. It discusses two main decomposition methods for generating such data - Eigen decomposition and Cholesky decomposition. These methods decompose the variance-covariance matrix in different ways to simulate random normal values that match the desired mean and variance-covariance structure. The document provides code examples in R to implement both methods and compares the results. It finds that both methods accurately reproduce the target multivariate normal distribution.
Application of panel data to the effect of five (5) world development indicat...Alexander Decker
This document discusses applying a panel data model to analyze the effect of 5 world development indicators (WDI) on GDP per capita for 20 African Union countries from 1981 to 2011. It introduces panel data modeling and the fixed effects model specifically. The fixed effects model is estimated using least squares dummy variable regression to account for country-specific effects. The results of analyzing the relationship between GDP per capita and the 5 WDI (exchange rate, money supply, inflation, natural resources, foreign investment) using this fixed effects panel data model are then presented.
Application of panel data to the effect of five (5) world development indicat...Alexander Decker
This document discusses the application of panel data analysis to examine the effect of 5 world development indicators (WDI) on GDP per capita for 20 African Union countries from 1981 to 2011. It presents the panel data model, describes the methodology used as fixed effects regression, and provides sample output of the panel data format and regression results. The key world development indicators examined are official exchange rate, broad money, inflation rate, total natural resources rents, and foreign direct investment.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Geoid height determination is one of the major problems of geodesy because usage of satellite
techniques in geodesy isgetting increasing. Geoid heights can be determined using different methods according
to the available data. Soft computing methods such as Fuzzy logic and neural networks became so popular that
they are used to solve many engineering problems. Fuzzy logic theory and later developments in uncertainty
assessment have enabled us to develop more precise models for our requirements. In this study, How to
construct the best fuzzy model is examined. For this purpose, three different data sets were taken and two
different kinds (two inpust one output and three inputs one output) fuzzy model were formed for the calculation
of geoid heights in Istanbul (Turkey). The Fuzzy models results of these were compared with geoid heights
obtained by GPS/levelling methods. The fuzzy approximation models were tested on the test points.
This document describes a stock price model based on the voter model from statistical physics. The model represents stock investors on a lattice who can be in a buying, selling, or neutral position based on interactions with neighboring investors. The stock price is then derived from the positions of investors. Computer simulations of the 1D model are presented, showing the fluctuations of generated stock prices and returns. The document also discusses properties found in real stock market data, such as power law distributions and volatility clustering, that the model aims to replicate.
This document describes an uncertain volatility model for pricing equity option trading strategies when the volatilities are uncertain. It uses the Black-Scholes Barenblatt equation developed by Avellaneda et al. to derive price bounds. The model is implemented in C++ using recombining trinomial trees to discretize the asset prices over time and space. The code computes the upper and lower price bounds by solving the Black-Scholes Barenblatt PDE using numerical techniques, with the volatility set based on the sign of the option gamma.
Regeneration of simple and complicated curves using Fourier seriesIJAEMSJORNAL
This document describes using Fourier series to regenerate simple and complex curves composed of sine waves. Fourier series is shown to be an effective method for this purpose. As an example, a periodic sine wave pattern with 22 data points is modeled using a Fourier series with 4 harmonics. The Fourier coefficients are estimated using a statistical software package. The estimated Fourier model is then used to regenerate the sine wave pattern with good accuracy compared to the original data points. The document demonstrates how Fourier series can reliably regenerate periodic patterns constructed from sine curves.
The document discusses the planted clique problem in graph theory. It introduces the problem and describes how previous research has found polynomial-time algorithms to solve the problem when the size of the planted clique k is O(√n). The document then summarizes two algorithms - Kucera's algorithm and the Low Degree Removal (LDR) algorithm - that have been used to approach the problem. It describes implementing the algorithms in a C++ program to simulate random graphs with planted cliques and test the ability of the algorithms to recover the planted clique.
A Strategic Model For Dynamic Traffic AssignmentKelly Taylor
This document proposes a strategic model for dynamic traffic assignment. The key elements are:
1) Users follow strategies that assign preference orders to outgoing arcs from each node based on arrival time and congestion.
2) A time-space network is constructed to model flow variations over time on the original road network.
3) An equilibrium is achieved when expected delays are minimal for each origin-destination pair given the strategies and capacities.
Non-life claims reserves using Dirichlet random environmentIJERA Editor
The purpose of this paper is to propose a stochastic extension of the Chain-Ladder model in a Dirichlet random
environment to calculate the provesions for disaster payement. We study Dirichlet processes centered around the
distribution of continuous-time stochastic processes such as a Brownian motion or a continuous time Markov
chain. We then consider the problem of parameter estimation for a Markov-switched geometric Brownian
motion (GBM) model. We assume that the prior distribution of the unobserved Markov chain driving by the
drift and volatility parameters of the GBM is a Dirichlet process. We propose an estimation method based on
Gibbs sampling.
Technology S-Curve Analysis (TSC) is a method to determine the relationship between investment in improving technology and corresponding market sales. The curve typically takes the shape of an S as the market develops through phases of launch, growth, maturity, and decline. TSC Analysis is particularly useful for launching new technologies and assessing the remaining lifespan of technologies to guide business strategy and shifting resources to next-generation technologies. TSC Analysis provides intelligence on the current technology lifespan, potential limits, economics of innovation, a company's position, and quantifying investment versus payoff.
The document discusses S curves, which plot cumulative project quantities like costs or hours against time. It describes different types of S curves, including baseline, target, and actual S curves. The document explains how S curves can be used to analyze a project's progress, growth, and slippage. It also provides details on generating S curves from project schedule data and interpreting S curve analyses.
The S curve is used to spread project costs over time in a bell-shaped curve. It is based on the sine and cosine waves, with the integral of the sine wave producing a leaning S-shaped cumulative cost curve. The document describes tweaking the basic S curve model by introducing variables to skew the timing of costs (make them front-loaded or back-loaded) and adjust the peakness to flatten or steepen the curve. Equations are provided to calculate cost at any point in time or rate of spending based on total project cost, time, skewness, and peakness variables.
The document provides safety instructions and an overview of the features and operation of the Dual 31 Band Equalizer. Key points include:
- The equalizer has 31 frequency bands covering 20Hz to 20kHz that can each be boosted or cut by up to 12dB or 6dB depending on the range setting.
- It has electronically balanced inputs and outputs, variable low cut filter, bypass switches, and LED VU meters.
- The document explains the controls on the front and rear panels and provides instructions for setting up and using the equalizer for different applications such as with a mixer, patchbay, or real time analyzer.
This document presents a case study analyzing the S-curve of a construction project. It describes the original planned S-curve, which projected slow initial progress accelerating in the middle before slowing at completion. Actual progress fell far below this, requiring three adjustments - adding scope, and extending the schedule by 9 months total - to get the project back on track. By the third adjustment, actual progress more closely aligned with the planned S-curve, allowing the project to progress through start-up, growth, and commissioning phases to completion. The case study demonstrates how S-curve analysis can identify issues and inform decisions to successfully manage a project over its lifecycle.
The document discusses optimizing S-curve velocity profiles for motion control. An S-curve velocity profile is a smooth curve that is differentiable to the second order. The document describes decomposing 3D motion into 1D components and synthesizing physical constraints on jerk, acceleration, and velocity for each axis. It then presents calculations for generating an optimized S-curve velocity profile that satisfies the constraints and produces smooth, fast, and accurate motion between a start and end point.
The document discusses how to generate S-curves in Oracle Primavera P6 to analyze project progress and performance. S-curves show cumulative costs, labor hours, or other metrics plotted against time and typically have an S-shape. In Primavera P6, S-curves can be generated by activity or resource in the usage profile windows. Various analysis can then be done by comparing baseline, target, and actual S-curves to determine project growth, slippage, and progress. The S-curves can also be published from Primavera P6 as prints or embedded in webpages.
This document describes a methodology for modeling S-curves to forecast cost distribution over time for construction projects. It discusses three approaches to S-curve modeling: 1) Analyzing cost-time curves from literature, 2) Examining data from completed projects, and 3) Creating standard critical path models. The results from each approach are presented through mathematical expressions and diagrams. Finally, the results of the three approaches are integrated to develop final S-curves showing standard cost dynamics over time for different project types (buildings, tunnels, motorways). The proposed methodology can be used to forecast cost schedules during early project phases when detailed information is limited.
This document summarizes an optimization model for a contractor's S-curve developed using genetic algorithms. The model aims to minimize total construction costs while considering the tradeoff between different resource productivity and costs of resource mobilization/demobilization. An example project is optimized, resulting in a smoother resource allocation and lower total cost compared to the early and late schedules. The optimal S-curve developed from the model provides a baseline for measuring impacts of changes on construction costs.
This document discusses innovation lifecycles and how companies can leverage S-curves to drive breakthrough growth. It contains the following key points:
1. Products, services, and technologies progress along S-curves over time from emergence to maturity. Understanding where opportunities fall in their lifecycles can help companies innovate and avoid disruption.
2. Companies often fail to transition to new S-curves due to not focusing on or defending emerging technologies, cultural inertia, or lack of foresight.
3. Organizational strategies like leadership, structure, and metrics should evolve to support innovation or optimization depending on a business's point in the lifecycle. Tailoring the organization maximizes growth across the portfolio
This document discusses exploring the limits of technology S-curves by examining their usefulness for managers in planning new technology development. It focuses on the disk drive industry as a case study. The author makes four key points: 1) S-curves accurately describe industry-level technology substitution patterns, 2) to improve products, managers must oversee both component and architectural technology development, 3) S-curves describe individual firm experiences with components but cannot prescribe strategy, and 4) attackers gain advantage in this industry through architectural, not just technological, innovation in new applications.
This document discusses the history and development of semiconductors and integrated circuits. It describes how the transistor enabled electronics to be performed using silicon, leading to solid-state electronics like transistor radios. The integrated circuit was developed using the planar process to fabricate multiple transistors on silicon wafers. Moore's Law, proposed in 1965, predicted that the number of transistors on an integrated circuit would double every 18 months. This prediction has proven remarkably accurate and has driven innovation in the semiconductor industry for over 40 years. Continued shrinking of circuit elements has enabled faster processing speeds, higher functionality, and lower costs over time.
The document summarizes the key points from the book "Jumping the S-Curve" by Paul Nunes and Tim Breene. It discusses how companies can succeed by repeatedly climbing business S-curves and jumping to new S-curves before performance plateaus. To climb an S-curve, companies must identify a large market opportunity, build the necessary capabilities, and attract top talent. To jump to a new S-curve, companies must manage hidden S-curves, develop edge-centric strategies, reconstitute leadership teams early, and cultivate a talent pipeline. High performers are able to continually reinvent themselves through repeated S-curve climbing and jumping.
This document summarizes a research paper on algorithms for planning s-curve motion profiles.
The paper generalizes the model of polynomial s-curve motion profiles in a recursive form. It then proposes a general algorithm to design s-curve trajectories in a time-optimal manner. The algorithm calculates the time periods for connecting trajectory segments to generate a smooth path that meets velocity and acceleration limits. Experimental results on a linear motor system demonstrate the effectiveness of the algorithms in generating s-curve motion profiles.
The document proposes an S-curve network model to describe finite networks with bulk growth. It summarizes that most network models assume infinite growth, but real networks are finite. The model adds new nodes exponentially at each time step based on a logistic curve, with the total number of nodes approaching a carrying capacity. It connects new nodes preferentially to existing high-degree nodes. The model aims to better represent features like the limited growth of real networks like the Chinese IPv4 address network.
This document analyzes commodity trade data between the US and Canada to test for the presence of an S-Curve dynamic. The author disaggregates bilateral trade data into 60 commodities representing over 80% of total trade. Cross-correlation analysis reveals support for the S-Curve pattern in the trade balances of 41 of the 60 commodities, providing evidence the effect exists at the disaggregated level where it was not found using aggregate data. Industries exhibiting the S-Curve include both durable and non-durable goods as well as large and small industries by trade share.
The document discusses performance analysis of digital signal processing (DSP) technologies using the S curve model. It begins by establishing the theoretical foundations of the S curve model, noting its relationships to innovation diffusion models and life cycle models. The S curve model represents the evolution of technology performance over time. The document then presents a methodology for applying the S curve model to analyze the technology performance of DSPs. Finally, it discusses the possibilities and usefulness of the S curve model as a strategic analysis tool.
Technology S-Curve Analysis (TSC) is a method to determine the relationship between investment in improving technology and corresponding market sales. The curve typically takes the shape of an S as the market develops through phases of launch, growth, maturity, and decline. TSC Analysis is particularly useful for launching new technologies and assessing the remaining lifespan of technologies to guide business strategy and shifting resources to next-generation technologies. TSC Analysis provides intelligence on the current technology lifespan, potential limits, economics of innovation, a company's position, and quantifying investment versus payoff.
This document discusses technology S-curves and their implications for renewable energy alternatives. It finds that both wind and geothermal energy are poised to become more economical than fossil fuels relatively soon based on historical cost reductions from R&D investment. However, government R&D funding has been lower for wind and geothermal compared to solar energy. The analysis suggests governments may over-fund fossil fuel technologies given their diminishing performance improvements. Examining renewable technologies through an S-curve lens provides insights not found using traditional experience curves and indicates areas where industry and government could better support technology development.
The document provides safety and operating instructions for the Dual 31 Band Equalizer. It describes the unit's features such as its dual channel design with 31 1/3 octave bands, constant Q circuitry, cut/boost ranges, and bypass switches. It also covers setup instructions, typical system configurations, equalization techniques, and specifications.
Accident detection system project report.pdfKamal Acharya
The Rapid growth of technology and infrastructure has made our lives easier. The
advent of technology has also increased the traffic hazards and the road accidents take place
frequently which causes huge loss of life and property because of the poor emergency facilities.
Many lives could have been saved if emergency service could get accident information and
reach in time. Our project will provide an optimum solution to this draw back. A piezo electric
sensor can be used as a crash or rollover detector of the vehicle during and after a crash. With
signals from a piezo electric sensor, a severe accident can be recognized. According to this
project when a vehicle meets with an accident immediately piezo electric sensor will detect the
signal or if a car rolls over. Then with the help of GSM module and GPS module, the location
will be sent to the emergency contact. Then after conforming the location necessary action will
be taken. If the person meets with a small accident or if there is no serious threat to anyone’s
life, then the alert message can be terminated by the driver by a switch provided in order to
avoid wasting the valuable time of the medical rescue team.
A high-Speed Communication System is based on the Design of a Bi-NoC Router, ...DharmaBanothu
The Network on Chip (NoC) has emerged as an effective
solution for intercommunication infrastructure within System on
Chip (SoC) designs, overcoming the limitations of traditional
methods that face significant bottlenecks. However, the complexity
of NoC design presents numerous challenges related to
performance metrics such as scalability, latency, power
consumption, and signal integrity. This project addresses the
issues within the router's memory unit and proposes an enhanced
memory structure. To achieve efficient data transfer, FIFO buffers
are implemented in distributed RAM and virtual channels for
FPGA-based NoC. The project introduces advanced FIFO-based
memory units within the NoC router, assessing their performance
in a Bi-directional NoC (Bi-NoC) configuration. The primary
objective is to reduce the router's workload while enhancing the
FIFO internal structure. To further improve data transfer speed,
a Bi-NoC with a self-configurable intercommunication channel is
suggested. Simulation and synthesis results demonstrate
guaranteed throughput, predictable latency, and equitable
network access, showing significant improvement over previous
designs
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
Digital Twins Computer Networking Paper Presentation.pptxaryanpankaj78
A Digital Twin in computer networking is a virtual representation of a physical network, used to simulate, analyze, and optimize network performance and reliability. It leverages real-time data to enhance network management, predict issues, and improve decision-making processes.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELijaia
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Open Channel Flow: fluid flow with a free surfaceIndrajeet sahu
Open Channel Flow: This topic focuses on fluid flow with a free surface, such as in rivers, canals, and drainage ditches. Key concepts include the classification of flow types (steady vs. unsteady, uniform vs. non-uniform), hydraulic radius, flow resistance, Manning's equation, critical flow conditions, and energy and momentum principles. It also covers flow measurement techniques, gradually varied flow analysis, and the design of open channels. Understanding these principles is vital for effective water resource management and engineering applications.
Impartiality as per ISO /IEC 17025:2017 StandardMuhammadJazib15
This document provides basic guidelines for imparitallity requirement of ISO 17025. It defines in detial how it is met and wiudhwdih jdhsjdhwudjwkdbjwkdddddddddddkkkkkkkkkkkkkkkkkkkkkkkwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwioiiiiiiiiiiiii uwwwwwwwwwwwwwwwwhe wiqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq gbbbbbbbbbbbbb owdjjjjjjjjjjjjjjjjjjjj widhi owqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq uwdhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhwqiiiiiiiiiiiiiiiiiiiiiiiiiiiiw0pooooojjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjj whhhhhhhhhhh wheeeeeeee wihieiiiiii wihe
e qqqqqqqqqqeuwiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiqw dddddddddd cccccccccccccccv s w c r
cdf cb bicbsad ishd d qwkbdwiur e wetwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww w
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddfffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffw
uuuuhhhhhhhhhhhhhhhhhhhhhhhhe qiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii iqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee qqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqccccccccccccccccccccccccccccccccccccccccccccccccccccccccccccc ccccccccccccccccccccccccccccccccccc bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbu uuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuum
m
m mmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm m i
g i dijsd sjdnsjd ndjajsdnnsa adjdnawddddddddddddd uw
Blood finder application project report (1).pdfKamal Acharya
Blood Finder is an emergency time app where a user can search for the blood banks as
well as the registered blood donors around Mumbai. This application also provide an
opportunity for the user of this application to become a registered donor for this user have
to enroll for the donor request from the application itself. If the admin wish to make user
a registered donor, with some of the formalities with the organization it can be done.
Specialization of this application is that the user will not have to register on sign-in for
searching the blood banks and blood donors it can be just done by installing the
application to the mobile.
The purpose of making this application is to save the user’s time for searching blood of
needed blood group during the time of the emergency.
This is an android application developed in Java and XML with the connectivity of
SQLite database. This application will provide most of basic functionality required for an
emergency time application. All the details of Blood banks and Blood donors are stored
in the database i.e. SQLite.
This application allowed the user to get all the information regarding blood banks and
blood donors such as Name, Number, Address, Blood Group, rather than searching it on
the different websites and wasting the precious time. This application is effective and
user friendly.
Road construction is not as easy as it seems to be, it includes various steps and it starts with its designing and
structure including the traffic volume consideration. Then base layer is done by bulldozers and levelers and after
base surface coating has to be done. For giving road a smooth surface with flexibility, Asphalt concrete is used.
Asphalt requires an aggregate sub base material layer, and then a base layer to be put into first place. Asphalt road
construction is formulated to support the heavy traffic load and climatic conditions. It is 100% recyclable and
saving non renewable natural resources.
With the advancement of technology, Asphalt technology gives assurance about the good drainage system and with
skid resistance it can be used where safety is necessary such as outsidethe schools.
The largest use of Asphalt is for making asphalt concrete for road surfaces. It is widely used in airports around the
world due to the sturdiness and ability to be repaired quickly, it is widely used for runways dedicated to aircraft
landing and taking off. Asphalt is normally stored and transported at 150’C or 300’F temperature
5G Radio Network Througput Problem Analysis HCIA.pdf
2007 02t
1. APPLICATION OF S-SHAPED CURVES
(edited transcript)
Presented at ETRIA TRIZ Future Conference 2007, Frankfurt
by Dmitry KUCHARAVY and Roland DE GUIO, 7th November 2007.
…A law that becomes violated is not much of a law1
…
The famous book about the innovator's dilemma [1] Cristinsen starts with insights from
the hard disk industry. It is interesting to notice, that those insights are based on the
rigorous data analysis and fitting logistic S-shaped curves to data. Some results of the
analysis were published 1992 in the article about limits of the technology S-curves [2].
Application of S-shaped curves is interesting but a narrow topic among many in our
research about Technological forecasting. The certain area of application for the
extrapolation methods by logistic S-curves for technological forecasting can be clarified
using publications of Ayres, Martino, Makridakis, Armstrong, Porter, Glen and others
[3, 4, 5].
The aim of presented results is to convince readers that it is dangerous practice to plot
S-curves as free-hand drawing. Unfortunately, it is usual practice, especially in TRIZ
society. Every second presentation within ETRIA TRIZ Future Conference 2007,
Frankfurt mentions S-curve, but all of them are done arbitrary by hand. Why is it
destructive? Despite such a practice provides quick and visionary results, free-hand
made curves mislead researches and hide the real trends.
We are going to provide our arguments, observed peculiarities, and conclusions through
six Questions about simple logistic curves:
Q1: What does S-curve mean?
Q2: Why does it work?
Q3: Where is it applied and why?
Q4: Qualitative or quantitative?
Q5: So what?
Q6: What would we do with it?
Q1: What does S-curve mean?
S-shaped curves are numerous and different not only by essence, but mostly by names.
They possessed many names2
through the history since at 1833 Belgian mathematician
Pierre-Francois Verhulst proposed logistic equation as model of population growth.
Basically all S-shaped curves can be subdivided by symmetrical and non-symmetrical.
In presented topic we are going to apply most generic name which reflects the very
essence: logistic curve of natural growth. Model of natural growth of autonomous
systems in competition might be described by logistic equation and symmetric (simple)
logistic S-curve where 'natural growth' is the tendency of parameters to increase its
value during time (i.e. evolution). While something growth, something is going to
1
From article of T.Modis
2
Logistic curve, Verhulst-Pearl equation, Pearl curve, Richard's curve, Generalized Logistic, Growth
curve, Gompertz curve, S-curve, S-shaped pattern, Saturation curve, Sigmoid(al) curve, Foster’s curve,
Bass model, and many other names can be found.
1
2. decline in accordance with First law of thermodynamics3
. Therefore, any growing
process is tightly linked with competition for resources. Many different types of
competitions are recognized [6, 7]; but this question is out of our scope.
In the scope of logistic equation the 'natural growth' is defined as the ability of a
'species' to multiply inside finite 'niche capacity' through a given time period. In other
words, for technological systems, 'species' are growing variable (variable the value of
which grows during time). For instance, efficiency of engine, population, size, weight,
number of words, … i.e. value we are interested about. In engineering practice, a
function of system and critical-to-quality features and their values are identified
explicitly. 'Niche capacity' for engineering systems can be interpreted as available
resources of space, time, substances, consumers and other material and non-material
(e.g. knowledge, cultural needs) elements required for operation.
Internal mechanism of natural growth under competition for resources can be presented
various ways. For instance in system dynamic one of the patterns of system's behavior
(system archetype) called 'limits to growth' is described with help of causal loop
diagram the next way:
Figure 1. Causal loop diagram for limits of growth4
Figure 2. Arbitrary diagram for limits of growth4
For socio-technical systems the three-parameter S-shaped logistic growth model is
applied for describing a continuous "trajectories" of system's growth or decline through
time.
βα
κ
−−
+
= t
e
tN
1
)(
(1)
Where, N(t) – number of 'species' or growing variable to study; e - the base of the
natural system of logarithms, having a numerical value of approximately 2.71828..
The three parameters that specify the curve, are κ, α, and β:
3
Thermodynamics: the first law of thermodynamics is a statement of the conservation of energy for
thermodynamic systems, and is the more encompassing version of the conservation of energy. The
conservation of energy states that the total amount of energy in any closed system remains constant but
can't be recreated, although it may change forms, e.g. friction turns kinetic energy into thermal energy.
(Wikipedia, Nov. 2007)
4
Source: Braun, W. The System Archetypes. The Systems Modeling Workbook. 2002.
2
3. κ – is the asymptotic limit of growth5
(it might depend on available space, market niche,
or carrying capacity); for case N(t) << κ, logistic model closely resembles exponential
growth;
α – growth rate parameter specifies "width" or "steepness" of S-curve;
it is frequently replaced with a variable that qualifies time required for "trajectory" to
grow from 10% to 90% of limit κ (carrying capacity), a period named characteristic
duration, or Δt;
β – parameter specifies the time (tm) when the curve reaches 0.5κ or the midpoint of the
growth trajectory N(tm)= κ/2; point of inflection tm implies symmetry of simple logistic
function;
Usually the parameters α and β are calculated by fitting the data.
Equation (1) is a solution of differential equation (2) applied as model for exponential
growth with limiting feedback:
−=
κ
α
)(
1)(
)( tN
tN
dt
tdN
(2)
Where limiting term
−
κ
)(
1
tN
is close to 1 if N(t) << κ and approach 0 as N(t) → κ.
In order to present each parameter explicitly, we apply fit to data for annual TRIZ-
publications in English from 1996 to 2006. Data collected on the basis of publications at
conferences within the time span 1999-2006 (ETRIA6
TFC 2001-2006, and TRIZCON7
)
and publications on the website triz-journal.com 1997-20068
. The same articles from
different sources where taken into account just once. Prepared data to fit logistic curve
are presented in Table 1.
Table 1. Cumulative number of TRIZ publications in English (1996-2006)
Year TRIZ Journal TRIZCON ETRIA TFC total, year cumulative
1996 5 5 5
1997 32 32 37
1998 70 70 107
1999 68 26 94 201
2000 65 32 97 298
2001 75 36 30 141 439
2002 86 41 51 178 617
2003 99 26 31 156 773
2004 87 27 49 163 936
2005 90 37 58 185 1121
2006 103 34 58 195 1316
Logistic curve is fitted to data using Loglet Lab version 1.1.4 software9
.
5
Another name is 'perceived potential'.
6
Annual TRIZ Future Conference of European TRIZ Association, Europe [http://www.etria.net ].
7
Annual Altshuller Institute for TRIZ Studies International Conference, USA [ http://www.aitriz.org ].
8
The TRIZ Journal Article Archive [ http://www.triz-journal.com ].
9
http://phe.rockefeller.edu/LogletLab/
3
4. Figure 3. TRIZ publications in English
How this fit is useful to forecast evolution of TRIZ publications can be checked next
years. The past data gives us opportunity to recognise a trend and to extrapolate it to the
future. We are going to apply this example to illustrate the three parameters of simple
logistic curve only.
In the bottom of Figure 3 the raw residuals are presented. Residuals show how
accurately the curve fits the data. Residual presents the difference between an observed
value and the fitted value for the same value of time. A correct fit has residuals
randomly above and below the x-axis. A cluster of consecutive points locating all
above or all below the x-axis may indicate an inaccurate fit.
Logistic S-curve represents cumulative growth of variable to study, when bell-shaped
curve is a symbol of growth rate. If we count number of TRIZ publications year after
year, the data will be close to the bell-shape distribution (so-called 'normal
distribution'10
). Data will represent a rate of growth. If the data will be represented as
cumulative number of TRIZ publications, the data can be approximated using simple
logistic S-curve. While the rate of grows follows a bell-shaped curve, the cumulative
growth traces out an S-curve. If we have bell-shaped curve, the S-curve can be derived
and vice versa (see Figure 4.)
10
Known as Gaussian distribution. Johann Carl Friedrich Gauss became associated with these
distributions after he analyzed astronomical data using them.
4
5. a) b)
Figure 4. Rate of growth and cumulative number of TRIZ publications
Different publications [8, 9, 10, 11] report higher performance of practical forecasting
with simple symmetrical S-curves. Thus, we focus our research efforts on application of
this type of curves, despite of existence of others [12]. The simple logistic is useful in
part because of its small number of parameters, the clear interpretation of obtained
results, it is easy-to-apply, and it is possible to apply for the description of more than
one logistic pulse (bi-logistic, and multiple logistics).
What are the meanings of each of three parameters for simple logistic curve? First
parameter is the ceiling of growth is represented by κ in equation 1. Proposed fit to data
results the expected value of κ (number of publications) within confidence intervals
(1430 < κ < 1750; Figure 7). Second parameter represents growth period. In TRIZ
dictionary it correspond to the time period between points α and β. This growth period
closely resemble to linear growth. Third parameter of simple logistic specifies time (tm),
when curve reach midpoint of the growth trajectory. This is point of symmetry of
simple S-curve. The tm corresponds to the vertex of bell-shaped curve.
It is necessary to notice, that fitting technique using past data gives opportunity to
obtain two parameters out of three as result of fitting curve to data. However, the point
of biases is the assumptions about upper limits of growth.
The results of study based on forty thousand fits [13] showed that the more precise the
data and the bigger the section of the S-curve they cover, the more accurately
parameters can be recovered. The reliability of forecasting about ceiling is higher when
available data cover more than half of S-curve. What can be done when there is lack of
data?
In scope of TRIZ, the causal method is applied to compensate the lack of data.
Frequently as causal variables are applied number of inventions, level of inventions and
profitability values [15]. Nevertheless, in his famous Creativity as an Exact Science
[15], G. Altshuller discusses about three levels of limiting resources a system faces
within its evolution along S-curve: limits of system resources, limits of available
resources and physical limits of resources in super-system.
To improve reliability of medium and long-term technological forecasting it is proposed
to apply the knowledge about limiting resources (causes), in order to forecast limits of
growth for particular system when lack of data for emerging technologies does not
allow applying the naïve method.
5
6. Naïve and Causal methods
Naïve methods: Yt-d, …, Yt-2, Yt-1, Yt
→ Yt+h
Causal methods: Xt-d, …, Xt-2, Xt-1, Xt
→ Xt+h
↓ b ↓ bh
Yt-d, …, Yt-2, Yt-1, Yt Yt+h
Figure 5. Naïve and Causal methods. Adapted from [14]
Where, Y – the variable to be forecasted; X – causal variables; d – periods of historical data; h –
periods in the forecast horizon; t – time period (e.g. year); b – the causal relationships in the
historical data; bh – the causal relationships in the forecast horizon.
Naive model: a model that assumes things will behave as they have in the past. In time series,
the naive model extends the latest observation. For instance, Random walk model (sub-set of
naïve models): it assumes that, from one period to the next, the original time series merely
takes a random "step" away from its last recorded position. Causal model: a model that goes
beyond of variable of interests by asking "why"?
Naïve methods apply past data about the variable to be forecast (Y) in order to identify the
trends and extrapolated them into the future. Causal methods apply causal variables (X) to
foresee future changes of target variable (Y). A causal variable (X) is one that is necessary or
sufficient for the occurrence of an event (Y). It is assumed also that X precedes Y in time. Past
data about causal variable (X) are used in order to identify trends and apply this knowledge to
foresee future values of target variable (Y).
Natural law of logistic growth under competition can be described by various ways. It
entirely depends on aims of analysis and selected values on the axes. For instance, the
logistic growth curve can be linearized with Fisher-Pry transform [16, 13, 9].
Figure 6. Fisher-Pry transform Figure 7. S-curve with confidence intervals
The plot on the Figure 6 presents the same data and trend as S-curve on the Figure 4.
Fisher-Pry transform facilitates comparison to other logistic growth processes, as soon
all the curves are normalized to limit of growth κ (see equation 3), more than one
logistic can be plotted on the same chart. Therefore, competition of multiple systems
over time can be simulated, presented and analyzed.
6
7. )
)(1
)(
()(
tF
tF
tFP
−
= , where
κ
)(
)(
tN
tF = (3)
When FP(t) is drawn on a semi-logarithmic scale, the S-curve becomes linear. The merit
of presentation using Fisher-Pry transform is ability to put on the same diagram
competitive systems with various performance and different working principles. The
"infant mortality" threshold11
corresponds to 10-1
(or 10%) on Figure 6. The "saturation"
threshold12
corresponds to 101
(or 90%) on Figure 6.
The diagram on the Figure 7 shows the results of fit to data with confidence intervals. In
accordance with applied algorithm to consider data accuracy and residual, the
cumulative number of articles (ceiling) lies in between 1435 and 1738 articles, when
midpoint of growing process is located between July 2002 and June 2003. Characteristic
duration of growth estimated in between 7.9 and 9.8 years.
In order to forecast technological systems evolution using natural law of growth and
logistic curves it is applied four assumptions:
1. How big will be value of limit of growth ("ceiling")? It is necessary to find a law
of nature (repeatable regularity) which will give proof of ceiling.
In order to improve accuracy of forecasting, it is proposed to analyze the
dynamic change of limiting resources for driving contradiction of system
evolution.
2. Logistic growth (S-shaped curve) can facilitate an accurate forecast. It is
proposed to apply the logistic model which is symmetric around the midpoint.
This assumption is based on hundreds of reported successful forecasts using S-
shaped curves and several reviews.
3. Characteristic duration (Δt) and time to reach midpoint of curve (tm) can be
defined using past data. In order to reinforce and recheck two out of three
parameters of S-curve it is proposed to apply knowledge about cycles of socio-
technical systems evolution.
[Kondratieff cycles, Schumpeter business cycles, Tchijevsky cycles, Mensch,
Marchetti, Modis, etc. ]
4. Technological evolution for certain function(s) is continuous process (e.g.
transportation systems). If the evolution from one technology to another one
took place in past, it will happen in future as well.
In order to improve reliability of forecast using logistic growth curves, it is necessary to
support each assumption by practical and theoretical results.
11
In TRIZ-literature it is named α-point.
12
In TRIZ-literature it is named β-point.
7
8. Q2: Why does it work?
…Everything can be fitted with enough parameters1…
Logistic curve describes evolution of system under limitation of resources through time.
In the other words, it represents changes of system parameters under competition. This
is the essential meaning of S-curve. What are the strong points and what are the weak
points of such representation?
Strength:
Properly established logistic growth reflects the action of a natural law. It works
everywhere, independently from scale. E.g. for nano-level (molecule clustering),
micro-level (yeast growth), macro-level (economy of country), and mega-levels
(stars and galactic growth).
Relatively easy to apply. Clear concept and working mechanism. It is clear what
is behind of plotted curve, those we trust to S-curve description.
Can be applied for systems where the growth mechanisms are understood and
where the mechanisms are hidden. For instance it is difficult to explain, why
number of articles for English TRIZ publication will not grow the same way as
during last five years. Presented results were unexpected before fitting logistic
curve to data.
…and Weakness:
What is growing variable (species) and what is the underlying competing
mechanism in particular case? It is not evident what can be selected to measure
growth of system evolution. Most of the presentations within the ETRIA TRIZ
Future conference have nothing on vertical axis. It is a challenging problem to
define growing parameter and to measure it. Lack of formal procedure to define
growing variable is not particular problem of TRIZ society. In border of our
research we tried to find the answer in different domains by asking practitioners
and reviewing literature. A formal procedure to identify growing variable with
confidence is 'on demand' till now.
Biases towards low or high ceiling: ‘no two people, working independently, will
ever get EXACTLY the same answer for an S-curve fit’.
Should we fit S-curve to the raw data or to cumulative number? This question is
not easy to answer especially at the beginning of forecasting. Which sort of data
is wholesome to disclose trends of particular system evolution? The answer
depends on essential mechanism of system growing under competition and how
do we perceive this mechanism.
Fitting technique errors and uncertainties mostly depends on applied data and
algorithm which have certain accuracy and precision.
Analysis of strengths and weaknesses of logistic S-curves application in context of
technological forecasting can be found at publications of Modis [11, 13], Ayers [10],
Christensen [2], Martino [4] and many others.
Q3: Where is it applied and why?
Before analyzing publications about logistic S-curve in English speaking TRIZ society,
it is interesting to see how many papers were published in one of the leading
8
9. international journal "Technological Forecasting and Social Change13
". Data about
publications at other journals: Future, International Journal of Forecasting, Journal of
Future Studies, Journal of Product Innovation Management, the Futurist, Ecological
Modelling and many others were not taken into account.
4%
53%
43%
Articles about peculiarities of S-curves application
Papers which mentioned S-curves
Papers which do not mentioned S-curves
May 2002 – May 2007 (320 articles)
Figure 8. Publications at the International Journal of Technological Forecasting and Social Change
(2002-2007).
The International Journal of Technological Forecasting and Social Change published
more than 320 articles from May 2002 to May 2007. About 14 articles considered in
particular the application of the logistic S-curve of natural growth for forecasting
purposes. More than 170 articles within this period mentioned S-curves on their pages.
Various books and publications report application of S-curves as a part of several
forecasting methods: Trend Impact Analysis, Curve fitting technique, Decision and
Statistical Modelling, Text and Data Mining, Life Cycle Analysis, Theory of Innovation
Diffusion, Emerging issues analysis and others [17, 18, 19, 20, 21, 22].
In order to predict technological futures for better decision making, studies about
technology maturity at the industry level are regularly performed. Our review of
literature showed, that most of them apply the logistic curves analysis as an important
part.
In his famous book 'Stalemate in Technology' [23], German scientist Gerhard Mensch
applied logistic curves to disclose regularities of basic innovation development for long
run of industrial revolutions. Logistic curves and Fisher-Pry transform give opportunity
to foresee technology substitution.
At the International Institute for Applied Systems Analysis (IIASA , Laxenburg, Austria
http://www.iiasa.ac.at/ ) logistic S-curves have been applied by Marchetti, Nakicenovic,
Grübler and other researchers during the last 35 years for studies about
the future of primary energy sources and vectors and energy dynamics,
the evolution of agricultural technologies,
the substitution of transportation systems,
the development of discoveries,
the elaboration of inventions and the diffusion of innovation,
the transformation of the aircraft industry,
13
http://www.elsevier.com/wps/find/journaldescription.cws_home/505740/description#description
9
10. macro- and micro- economic trends,
the growth of crime and terrorism,
environmental changes and problems,
the evolution of telecommunication systems...
Detailed studies about technology maturity and substitution for different industries were
done by Constant (1980) for aircraft engines; Roussel ( 1984) for foam rubber; Grübler
(1987) for steel and coal industries; Modis and Debecker (1988) for computer industry;
Van Wyk, Haour, and Japp (1991) for permanent magnets; Christensen (1992) for disk
drive industry; Ausubel and Marchetti (2001) for transport systems; Modis (2005) for
Internet growth; Foster (1986) for many examples from a range of industries. It is
impossible to take into account whole set of studies done with use of logistic curve as
soon it is a common practice during last decades.
The main question for such studies is about technological substitution: When and How
one technology will substitute another one? For instance, within period from 1980 to
1990 more than 50 research reports in English were issued, dozen published books
present hundreds diagrams with logistic S-curves. In his article about strengths and
weaknesses of S-curves [11] Theodor Modis write: “For the last 22 years I have been
fitting logistic S-curves to data points of historical time series at an average rate of
about 2–3 per day. This amounts to something between 15,000 and 20,000 fits.”
It is difficult to find some area of systems evolution were S-curves analysis has not been
applied yet. Hundreds of biographies of creative persons like Mozart, Bach, van Gogh
were presented through logistic curves. In order to understand some mechanisms of
scientific discoveries the process of discovery the stable chemical elements was
analysed as sequence of successive S-curves (Figure 9).
Figure 9. Clusters of discovery chemical elements. Source: [13]
Study using logistic curves about virus propagation and competition among death
diseases leaded to unexpected conclusion that "…disease starts to phasing out well
before effective medication becomes perfected and widespread…" [13].
Computer technologies dynamics and transfer of market value from platform
IBM+DEC to the platform Microsoft+Intel follows logistic curve as well (Figure 10).
The midpoint of two S-curves, when two platforms shared marked equally correspond
to 1993.
10
11. Figure 10. Transfer of market value. Source: [13]
Competition between two technologies can be presented in linear vertical scale.
However, for competition between several technologies it is much more practical to use
the logarithmic scale (e.g. see Figure 6).
International Institute for Applied Systems Analysis is a nongovernmental,
multidisciplinary, international research institution. It was founded in October 1972 by
academies of science of 12 nations from East and West. Now, the institute has 18
National full Member Organizations (http://www.iiasa.ac.at/ ). This unique institution,
established in time of Cold War with participation of United State, former Soviet Union,
Germany, Japan and other countries. What is the link between this multidisciplinary
research institution and logistic curve of growth? During more than 35 years, researches
of this institute apply the law of logistic growth for their research. Incomplete list of
projects covers next areas and technologies:
• the future of primary energy sources and vectors,
• the evolution of agricultural technologies,
• the substitution of transportation systems,
• the development of discoveries,
• the elaboration of inventions and the diffusion of innovation,
• the transformation of the aircraft industry,
• macro- and micro- economic trends,
• the growth of crime and terrorism,
• environmental changes and problems,
• the evolution of telecommunication systems…
Interim reports and many publications about are available through the Internet (e.g.
http://www.iiasa.ac.at/Admin/PUB/ , http://www.cesaremarchetti.org/publist.php )
today.
For instance, how does a long term study about primary energy sources look like?
11
12. Figure 11. Competition between primary energy sources.
Source: Adapted by Modis T. [13] from a graph at [24]14
.
Dots on the Figure 11 present data when lines are the fitting results. We can estimate
how good lines correlate with the data. In order to understand mismatching of data and
curves it is necessary to notice, that vertical scale is logarithmic.
What is fascinating at Figure 11? It is unknown what it will be15
, however it is projected
when and where the 'unknown' primary energy source will start to grow and what will
be the speed of growth.
On the Figure 11, at each moment there is 100% of energy represented through different
primary energy sources. For instance, the substitution process shows that the major
energy source between 1880 and 1950 was coal, when oil became the dominant source
from 1940.
Q4: Qualitative or quantitative?
The intriguing question is: should we perform qualitative study in favor of quantitative
one? Regarding to TRIZ traditions the qualitative analysis using S-curve of system life
cycle is preferable for inventive problem solving practice [26]. Outside of TRIZ
researches society there are proposals for applying S-curves qualitatively as well. For
instance it is suggested by Molitor to develop "patterns of change" based on qualitative
application of S-curves [22]. It was reported about 100 pattern of change which can be
applied in border of forecasting models.
However, there are much more case examples of quantitative application of logistic
curves of growth application. More than 150 articles of Marchetti, dozens articles and 4
books from Modis, dozens articles and software from research team of Ausubel are just
a part of the long list for quantitative application of logistic S-curves for study about
artificial and natural systems.
14
"Data, fits, and projections for the shares of different primary energy sources consumed worldwide. For
nuclear, the dotted straight line is not a fit but a trajectory suggested by analogy…The small circles show
how things evolved since 1982 when this graph was first put together" [13, p.161]
15
"… The futuristic source labelled “Solfus” may involve solar energy and thermonuclear fusion…" [13,
p.162]
12
13. Using analysis, proposed by Altshuller [15] it is possible to identify a position of
analyzed system on the S-curve. Even the questions regarding to vertical axis of
diagram and exact plot of S-curve itself stay unclear. This approach was adapted and
tested in different countries for different technologies [27, 28, 29, 30, 31, 32]. Presented
studies show the potential and limitations of analyzed technologies qualitatively using
quantitative analysis of inventions dynamic, level of inventions and profitability.
Nevertheless, such analysis has not been applied for study about competition of several
technologies. In fact, it does not present the answer for questions When and Where the
technology "B" will substitute the technology "A"?
Profitability
Levelofinventions
#ofinventions???
a)
Unit sales of U.S. music recording media.
Fitted logistic substitution of U.S. music recording media.
b)
Figure 12. a) quantitative TRIZ application of S-curve; b)qualitative analysis: US music
recording media (1997). Source: Adapted from [25]
On the Figure 12 b an analysis of US music recording media using the quantitative
logistic substitution model is presented [25]. In a time, when forecast about U.S.
recording media was done, it was unknown, what will be the technology to conquer CD
disks. Authors supposed it would be DVD. However, despite mistaken view about
technology, quantitative forecast using logistic substitution model predicted pretty well
the time and the speed of a new technology (MP3) growth.
The answer for question about "qualitative or quantitative" is not so easy. Now we can
not answer for the question directly. Both approaches have strong and weak points.
Before reformulating the question it is necessary to look to the nearest super-system
(where forecast suppose to be applied): What is more useful for decision making16
and
decision taking? Results of performed research about the last question can be expressed
in a shape of the following contradiction:
16
Decision making refers more to development decision and prepare all necessary components for,
whereas decision taking is the act of deciding. Decision taking results responsibility.
13
14. If method to assess technology barriers apply qualitative analysis, then it can be applied for long
term forecast due to conformity with law (of dialectic) of transformation quantity to quality;
however, it is difficult to achieve repeatable results from experts, it costs a lot, it takes a lot of
time (low frequency to update results), the results contains a lot of biases.
If method to assess technology barriers apply quantitative analysis, then the results can be
obtained a repeatable way, the process is cost effective, it is possible to update result frequently,
the results consist less biases; however, it is not compatible with law of transformation 'quantity
to quality', consequently it is mostly applied for short-term forecast.
It is required to forecast the sequence of transformations quantity to quality and to be
able to measure new (unknown at present time) quality in order to apply quantitative
models for continuing the chain of technology transformations.
law of transformation quantity to quality
"For our purpose, we could express this by saying that in nature, in a manner exactly fixed for
each individual case, qualitative changes can only occur by the quantitative addition or
subtraction of matter or motion (so-called energy)."
[Engels' Dialectic of Nature. II. Dialectics. 1883]
In other words, for long term forecast, it is necessary to answer: How can we satisfy the
law (dialectic) of transformation quantity to quality in a long run?
In fact, the analysis should be qualitative and quantitative. The desired result about the
method to study the future can be formulated the following way: it provides repeatable
results, using cost effective process, results can be updated frequently, results do not
consist biases of experts, it represent transformation of quantity to quality, it describes
future qualities a measurable way.
In order to summarize Pro and Con of qualitative and quantitative application of S-
curves the table on Figure 13 is proposed. This figure presents results of a brief
comparison analysis of application the logistic S-curves in three contexts: inventive
problems solving, management of innovations and technological forecast.
For instance, in border of inventive problem solving context, one of the critical question
for quantitative application of S-curve analysis is: How to measure a new quality?
Pioneering inventions (Altshuller, 1979) or Basic innovations (Mensch, 1979) produce
new qualities. Measurement of anything implies comparison with known things.
Cognition question arises in such a situation. Probably due to this issue, the basic
innovations take so long time to be adapted. For instance, according to Mensch's study,
it took 111 years for photography, 20 years for vulcanized rubber, 26 years for gasoline
motor, 79 years for incandescent light bulb, 29 years for high voltage generator.
14
15. Context Qualitative Quantitative
Inventive
problem solving
Pro: Quick and easy to perform. No
necessity to collect and refine data.
Con: Ambiguity of definition (e.g.
large, small). Partial & biased.
Pro: Clear definition of
features and values to improve.
Con: How to measure a new
quality?
Management of
innovation
Pro: Low resistance for
implementation.
Con: How to position the innovations
in time, in space, and in competitive
environment?
Pro: Plausibility for decision
making and strategic planning.
Higher-impartial.
Con: It takes a lot of efforts for
data gathering and refining.
Technological
forecasting
Pro: Compatible with long-term
forecast.
Con: Inaccuracy of forecasting in time
(when?) and in space (where?).
Higher-biased. How to deal with
competitive technologies?
Pro: Results are measurable
and precise. Repeatable,
adaptable, and cost effective.
Con: Based on past data and
trends. Indirect biases through
computation models,
assumptions and data.
Figure 13. Qualitative or Quantitative application of S-curve study.
Other examples are from technological forecasting context. First, How to take into
account by qualitative S-curve analysis the competitive technologies? It is generally
known, that any system evolves in the competitive environment. How to consider this
fact in border of qualitative analysis of S-curves?
Second, quantitative analysis using logistic S-curves can hide biases and assumptions in
the applied computational models. That is why the simple (symmetric) logistic S-curves
are so popular for forecasting. Thanks to its simplicity, it does not give a place for
hiding biases and preconceptions. For complex mathematical models it is difficult to
differentiate biases and model peculiarities.
Third, based on the past data, quantitative logistic curves reveal the trends from past.
This is a reason they have a limited foresight. New trends appear as synergy of
competition and cooperation between technologies permanently. How to incorporate
them in the forecast models?
In TRIZ society, the dominant S-curve analysis is qualitative one. From 1996 to 2006 it
was published more than 1300 articles in English language. In order to summarize the
application of S-curves in the TRIZ community we made an express survey of
publications at conferences within the time span 1999-2006 (ETRIA17
TFC 2001-2006,
and TRIZCON18
) and publications on the website TRIZ Journal19
(1997-2006). The
same articles from different sources where taken into account just once. Unfortunately,
we had no opportunity to consider publications from MATRIZ conferences and many
other publications in Russian, German, French, Spanish, Japanese, Korean, Chinese,
Polish, Czech and others languages.
17
Annual TRIZ Future Conference of European TRIZ Association, Europe [http://www.etria.net ].
18
Annual Altshuller Institute for TRIZ Studies International Conference, USA [ http://www.aitriz.org ].
19
The TRIZ Journal Article Archive [ http://www.triz-journal.com ].
15
16. Out of 1300 publications only 11 papers present the case studies with trials to apply S-
curves quantitatively. Total number of papers which mention S-curve of technical
system evolution is about 137 (see Figure 14). The diagram on the Figure 14 shows the
dynamic of interests in application of S-curve in different contexts in three periods:
before 2000, from 2000 to 2002, and from 2003 to 2006. In context of forecasting there
were published about 20 papers about S-curve usability. In context of problems solving,
the growing interest to S-curve outputs 17 papers. Case studies with quantitative
analysis about patents, profit and level of inventions result 11 publications. Most of the
articles just mention S-curve as useful analytical tool without details – 83 papers.
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
Non-grouped hypothesis 0 3 3
In context offorecasting 4 8 8
In context ofproblem solving 3 6 8
Case studies using S-curve 3 3 5
Just mentioned 11 29 43
1996-1999 2000-2002 2003-2006
Figure 14. TRIZ-publications about S-curve in English.
Q5: So what?
It is necessary to notice that:
• Extrapolation of trends using logistic S-curves model is essentially quantitative
forecasting methods with qualitative interpretation of results.
• When the analyzed process cannot be measured a qualitative application of S-
curves can produce misleading results and conclusions.
• The forecast can be dramatically different, depending on the selected parameters
and the way to scale it using simple logistic curves.
Why is it dangerous to plot S-curve by hand without clear parameters on the vertical
axis? It is misleading to plot S-curve without measurable parameters as soon we put our
unverified assumptions and hypothesis, but present them as the results of data
observation. Non-existing, imaginable trends are presented as result of practical
verification with data. In fact, such "trends" are not the trends, but only ideas how trends
can be. Instead of observation of facts and tendencies based on past data, there are
hypothetical trends based on particular experience, private opinions and speculation.
Such a practice reinforces biases and complicate situation in scope of long-term
forecasting when many competitive technologies and several contexts should be taken
into account. Such a practice leads to time and other resources consumption for going
wrong directions.
16
17. Self-complication and self-confusion take place as result of drawing S-curves by hand,
without measured data behind. Obtained "trends", based on the particular hypothesis
can not be validated, as soon there is no data to be checked and no algorithms to
construct curves behind of them.
In forecasting practice it is not enough to arrange measurable growing variable. It is
important to scale growing variable properly. Wrong scaling of speed as growing
variable leaded to forecast that 'the speed of light seem to be achieved by 1982' (left plot
on Figure 15). Using the same historical data and "envelop curve" as technique the
forecast at the right plot of Figure 15 is more reliable. Plotted on a different scale these
diagrams lead to a different sort of prediction.
Figure 15. Two ways to scale growing variable. Source [3, p.21]. The forecast can be dramatically
different, depending on the selected parameter for measuring the evolution of the system.
Figure 15 is taken from the book [3] about technological forecasting which G.Altshuller
cited when he wrote his manuscript [26]. The issues linked with growing variable
measurement and allocations are known a long time ago. Nevertheless, practice to draw
S-curve by hand is in the habit of many researches until now. It is interesting to
understand why so?
Extrapolation of trends using logistic S-curves is quantitative forecasting method based
on data about past of system transformation. In practice of technological forecasting S-
curves are used qualitatively to obtain rare insights and intuitive understanding of future
changes. Thus, it is recommended to start with fitting S-curves on data and continue
with qualitative analysis and interpretations.
Q6: What would we do with it?
The most interesting question we are working on is "How to forecast the future of
emerging technologies using simple logistic S-curves, when there is no statistical data
about?" In other words, "How to plot S-curve before system pass 'infant mortality
threshold' (point α at Figure 7)?" When system passes α point of its evolution some data
can be collected. There is still a question of data availability and the choice of an
appropriate growing variable. However, before 'infant mortality threshold' there is no
17
18. statistical data at all, because system does not exist out of laboratories. We are working
on question "How to construct the logistic S-curve before having statistical data for
growing variable trend?"
It was proposed to use causal method for approaching the question above. In the first
experiment (2004-2005: project about forecasting the small fuel cell technologies) it
was tested the hypothesis about studying problems (contradictions) as causal variable to
foresee future evolution of technology [33]. Number of contradictions was applied as
measurement to judge about technology maturity. This hypothesis was tested second
time (2005-2006) in border of project about future of distributed energy generation
technologies.
Nowadays (2006 - at present), we try to test an extension of original concept of
'contradictions as causal variables'. There are two basic assumptions behind: 1) any
process can be considered as learning process [13] especially problem solving; 2) at the
output of any learning process, there is knowledge acquisition issue. Therefore, it is
proposed to measure the knowledge growth within transition from invention to
innovation. In his book [23] Mensch presented data about 113 basic innovation history.
The distance between invention (feasible prototype) and innovation (first production for
market) for different technologies was different. For instance, for photography it was
111 years, for electricity production - 92 years, for dynamite - 23 years, for magnetic
tape recording - 39 years, and for fluorescent lightning - 82 years.
The research question can be reformulated: How to foresee time, place and peculiarity
of transition from invention to innovation in advance? The questions about forecasting
of inventions and the mature technologies time of retirement lay out of scope of this
description.
In order to better understand the main function of technological forecast the
photography example can be considered. Photography was invented in 1727. Let us
imagine that we are at 1790. The technological forecast question can be: When and
where the photography will be commercialized? In other words: How long will be the
distance from invention to innovation for emerging technology? (It was introduced as
market product 48 years later in 1838).
According to result of research in organization and economic sciences, the transition
from invention to innovation follows three consecutive stages [34, 35]: exploration
(research in laboratory), experimentation (field tests) and exploitation
(commercialization). The research assumption is: if to measure knowledge growth
during these stages, it is possible to predict with logistic S-curves the beginning of
commercialization (when system passes the α-point on its curve of evolution). It is
proposed to apply as growing variable the amount of knowledge about emerging
technology.
18
19. 100
0
50
25
75
Time
Exploration
(invention)
Exploitation
(innovation)
Experimentation
(Field test)
Invention Innovation
Knowledge acquisition,
in percents of readiness to
transit to the next stage
Figure 16. Growth of knowledge during exploration, experimentation and exploitation.
In order to manage fits of logistic curve, it is unavoidable to define the measurement
units for knowledge. Despite many publications in artificial intelligence, economics,
cognitive science, and pedagogy we did not find a relevant answer for question: "How
to measure knowledge?" This question seems a stumbling block for conducted research.
In fact, there are several feasible concepts how to measure information in particular
cases [36, 37], but the question about knowledge measurement is still the open one.
It is proposed to apply the network of contradictions as a guideline for the process of
knowledge acquisition. Among many others roles, the network of contradictions helps
to differentiate signal and noise on the early stages of emerging technology
development.
As a working concept it is adapted the relative ratio 'knowledge acquisition in percent
of readiness to transition to the next stage' as an interim answer for measurement units
(see Figure 16). One hundred percent represents sealing of knowledge acquisition for
certain stage (e.g. exploration). Right dashed curve represent a new system at
exploration stage. It is suggested that knowledge are accumulated during a time and
when there is enough knowledge to decide about next stage, S-curve of next stage (e.g.
experimentation) passes through α-point. It is supposed that when accumulated
knowledge approach 90% of limit of growth, it is the time to take decision about
transition to the next stage.
In practice, the experimentation (field test) stage can be launched far before exploration
stage of knowledge acquisition reach 90% of saturation. A weak point of proposed
assumption is fuzziness about when a new curve of knowledge growth substitutes the
old one and how to distinguish these two curves. Nevertheless, example about clusters
of discovery the chemical elements (Figure 9) and methodological advancement about
bi-logistic growth models [9] provide us confidence about future results.
If we take into account that all technologies evolve under competition it becomes clear,
why certain inventions will never reach the experimentation stage, when some of the
inventions that pass through 70% of experimentation will not arrive at exploitation
phase [35]. Reliable technological forecast should provide an explicit answer for the
question which (?) technology will win a competition, when (?) it will happen and
where (?).
19
20. The proposed working hypothesis about mechanism of knowledge acquisition is
suggestive one and it should be checked through practice. It is supposed that at stage of
exploration (E1), most of the knowledge is implicit at the beginning. At the end of the
stage E1 knowledge are represented mostly in scientific papers and patents. At the
experimentation stage (E2), needs in explicit knowledge increase, however it is
necessary to protect the intellectual properties. Therefore, most of the knowledge at the
beginning of E2 can be found in internal reports about field tests, in review, and in local
patents. At the end of the stage E2, the international patents, publications in industrial
journals increase in number, conference papers, and marketing articles are numerous.
* * *
What is proposed as working hypothesis for testing in coming future:
1. to measure knowledge growth by applying a network of contradictions as a
guideline to differentiate signal and noise information;
2. to employ the concept of limiting resources from super-system for validation the
network of contradiction for system;
3. to adapt the knowledge growth factor as an underlying cause of technology
substitution mechanism.
1. Signal and noise information can be differentiated when one focuses its attention not
on the existing technological solutions, but on the problems to be solved regardless to
known answers. Network of contradictions is a technology to realize the basic principles
of system thinking: "First, one should examine objectives before considering ways of
solving a problem. Second, one should begin by describing a system in general terms
before proceeding to the specific. [14]"
2. Application of simple logistic S-curves to represent growth of knowledge follows the
same concept of 'limiting resources' from nearest super-system as it was implemented to
study the evolution of technical systems. For instance, there is well known issue when
at the certain stage, the new laboratory experiments do not provide additional
knowledge about research topic. A typical answer for such an issue is to redesign
experiments or to conduct a field tests in real conditions but not in laboratory. There is
an open question for us what are the limiting resources in proposed example.
Analysis of limiting resources for constructed network of contradictions helps to review
and to validate obtained map of problems through study how formulated problems are
recognized in research and development societies. In the same time, study about
limiting resources discloses future problems and technological barriers.
3. According to preliminary results of our research, knowledge growth mechanism is
one of the major factors in the chain of technology substitution issues. The competition
issue is the exterior side of technology substitution when knowledge acquisition is an
internal force for surviving under competition.
What has been achieved?
The results of observation is based on 93 references20
: 4 books, 6 books sections, 10
conference papers, 13 official reports, 9 interim research reports, 51 journal articles
from Futures, International Journal of Forecasting, Technological Forecasting and
20
In addition to the titles from TRIZ community.
20
21. Social Change, and others. The performed study shows that the simple logistic S-curve
has been used extensively in the wide range of application. It produces appropriate
outputs for technological forecast of large spectrum of technologies.
The working assumption we are trying to check in scope of research about
methods for technological forecasting with logistic S-curves is: "efficiency of
knowledge growth determines technology substitutions and technological
competition mechanisms."
The basic working hypothesis to be checked: emerging technologies future can be
reliably predicted by applying simple logistic S-curves of knowledge growth in
framework of transition from invention to innovation21
.
How do we suppose to verify and validate the suggested hypothesis? Two times they
were tested through forecasting projects with European Institute for Energy Research
(EIFER, Karlsruhe). Interim results were reported. Three conference papers are
presented and communicated. Developing approaches, techniques and method will be
checked through practical forecast projects.
The ultimate test of the forecaster is an accurate and reliable forecast not the elegant or
easily applied method. Theodor Modis, 2007
21
Consecutive chain of exploration-experimentation-exploitation.
21
22. Questions:
Nikolai Khomenko:
You presented that network of contradiction can be a tool to distinguish the difference
between useful information and useless information. Right? Could you explain a little
bit more, how it can be done?
Dmitry Kucharavy (DK):
We discussed yesterday [5] about this question. Thus, I would like to open some slides
from yesterday tutorial to be constructive and short.
We construct a network of contradictions in order to have a guideline and to identify the
appropriate information (appropriate to formulated problems). At the beginning many
drawbacks seems like problems. This is the issue of 'perceived problems'. At the
beginning of study, specialists say "we have plenty of problems". When we represent
those 'problems' in form of contradictions many 'problems' disappear as soon it was just
naming the same things by different words, or situation was perceived as problem due
to the lack of information and communication between specialists. Afterwards, when we
try to put formulated contradiction into the map of contradictions (in OTSM-TRIZ
dictionary such form of contradictions is named: contradictions of parameters). Some of
previously formulated contradictions disappear (through its inconsistency,
reformulation, and integration with others contradictions) and new unobvious
contradiction appear in order to fulfill the map.
These observed results lead us to the conclusion that map of contradictions can be
useful for judging which information is appropriate and which one is not suitable for
objective of forecasting.
See an example of contradiction map on the Figure 17. This map is based on the
measurable critical-to-X features. These features were defined on the previous stages of
forecasting. All elements which are the sources of problems are linked with critical-to-X
features through couples of opposite requirements. For instance, Stack - Pressure inside
has to be high in order to satisfy 3. Electrical Efficiency, but Stack - Pressure inside has
to be low in order to satisfy 5. Maintenance intervals.
22
23. Figure 17. An example of map of contradiction for low temperature small fuel cell technologies.
Source: [33] (Information provided courtesy of EIFER, 2005, Karlsruhe)
As soon the problems are defined in context of map, they can be applied to monitor
growth of knowledge how this problem is solved. In other words, how future concept
solutions and technologies answer for the formulated problem.
The final map (see Figure 18) represents the same set of contradiction with additional
time axis. This axis represents integral estimations about speed of knowledge growth for
identified research and development activities. Research and development activities
were identified through monitoring of publications and reports. In order to select
publications, the map of contradictions was applied. Monitoring of all unselected
publications about small stationary fuel cell needs enormous human resources and time.
There should be criteria to select relevant information (distinguish signal and noise).
The mapped problems play the role of criteria to select relevant information.
When identified R&D activities were not linked with formulated problems, two variants
were considered: either we missed a problem on the map, either we have irrelevant
information. After several working sessions the map of contradictions was stabilized so
well that we did not observe issue about missed problems on the map.
Nikolai Khomenko:
When you developed the network of contradiction you are talking about value and
parameters. It is a kind of equivalent of physical contradiction. What is the contradiction
of parameters22
?
DK:
Speaking in terms of the dictionary of classical TRIZ, we constructed the map of
physical contradictions. However, as soon we face with engineering and non-
engineering contradictions during forecasting, the dictionary of OTSM-TRIZ is more
suitable: such contradictions are named as 'contradictions of parameters'.
22
The question on the record was not recognized exactly.
23
24. Figure 18. Map of contradiction for low temperature small fuel cell technologies with time scale.
Source: [33] (Information provided courtesy of EIFER, 2005, Karlsruhe)
Question:
You presented S-curve with upper limitations about TRIZ-publications in English. How
did you define the upper limits?23
DK:
The presented prediction about upper limits is based on the time series data from history
of last 10 years publications (see Figure 7). The data about dynamic of publications in
combination with simple logistic growth model [9] produces a particular shape of S-
curve. In other words, fitting logistic to time series data using non-linear least-squares
regression algorithm (Loglet Lab version 1.1.4 software) with minimum residuals
outputs values of the three parameters of simple logistic S-curve. One among others is
the upper limit κ (the asymptotic limit of growth).
When we look to the data with naked eyes, it is unobvious what will be the trend of
publication in coming future. Assumption about logistic nature of growth process and
the data from history of publication in combination with fitting technique disclose
future upper limit of growth. First we try to fit S-curve to data. Second, using
bootstrapping technique we try to minimize the residuals (i.e. errors - difference
between S-curve and real data). In example with TRIZ-publications in English we
applied so called 'naïve' method of forecasting (See Figure 5).
If to pose the question "What are the causes of publications dynamics?" - and to collect
data about causes, another curves can be plotted using 'causal' method. Presented curve
and the upper limit of TRIZ-publication in English has been obtained by 'naïve' method
using fitting the simple logistic to time series data.
Comment: Using symmetric logistic curve…
DK: Yes. Right.
Why do we apply the symmetric logistic? Most of the reviewed reports and articles [8,
10, 11, 12, 19] stated that the simplest methods perform more reliable forecast in
23
The question on the record was not recognized exactly.
24
25. comparing with complex and sophisticated mathematical models. In one of his
interview INSEAD24
Dean of Faculty and Professor of Decision Sciences, Anil Gaba
summarized "complicated complex models don’t actually predict real-life events" [38].
His colleague Spyros Makridakis, who spent his entire career looking at forecasting, is
sure that "simple methods do at least as well as some of most complicated methods".
Another reason to apply the simple logistic S-curve was the availability of free software
from Program for the Human Environment ( http://phe.rockefeller.edu/LogletLab/ ).
The Loglet Lab is a software package for analyzing logistic behavior in time-series data.
Instead of complicated non-symmetric logistic the researches from Rockefeller
University under leadership of Jesse H. Ausubel proposed to apply bi-logistics and
multiple-logistics to describe systems with two or more growth phases.
Gaetano Cascini:
When you present the evolution in number you measured (e.g. number of published
papers) you make it in time series (e.i. how many was published every year). How do
you identify the intervals that will allow you to forecast? Why not to choose number of
publications per month? Why do you apply 10 years period, but not 5 years period?
Probably, the result of forecast will be different.
Of course, when you increase time delay, you reduce fluctuations (may be) and,
probably the shape of the curve will be modified. Are there any suggestion for this
issue?
DK:
We did this kind of experimentation for our curiosity. It is correct, the fluctuation
increases for shorter time periods. However, the overall shape of fitted S-curve is the
same. Even more, we tried to plot logistic curves separately for TRIZ Journal,
TRIZCON conferences, and for ETRIA TRIZ future conferences. The obtained
forecasts were similar. When we tried to fit bi-logistic or multiple-logistic to the data, it
caused larger residuals (i.e. incorrect fit). In fact, during fitting logistic S-curve to data
we work for reduction the distance (error) between the real data and logistic curve. All
the forecasting power of S-curve study is based on the assumption that logistic curve
represent the natural growth law.
Gaetano Cascini:
The second question is: When we analyze some data (e.g. number of patents in a
particular domain) during a time, the growth process, as a rule, does not resemble the S-
curve. My suggestion is: there is a synergy of different S-curves and other patterns of
evolution. That is why the resulting curve to fit data has not S-shape. If we are talking
about logarithmic curves, we cannot apply the superposition principle.
Comment DK: We are talking about logistic curves. However we can plot them in
logarithmic scale as linear curves.
Gaetano Cascini: This is exactly what I was saying. If it is not true, just putting
together different curves you think everything can be curved25
.
DK:
There are several questions can be recognized in your 'one question'.
24
One of the world’s leading and largest graduate business schools
http://www.insead.edu/discover_insead/who_we_are/quick_facts.cfm
25
The question on the record was not recognized exactly.
25
26. I do agree with your suggestion, that it is really an art to identify what is a growing
variable for the question we would like to forecast. It is not easy at all. This is one of the
points we need to find a formal procedure to deal with. This is one of the major causes
why do we mix together frequently parameters that growth with different rhythms in
accordance with different patterns. As result we obtain a misleading forecast.
On the other hand, we disagree about different patterns for law of growth (in original
speech it was: "growing based on data"). Our observation and cognitive capacities are
limited as soon we are human being. When we look at some data, depending on their
nature (number of elements and so on) we do not agree a priory that it can be described
by S-curve. However, when we try to fit curve to data (wow!) it fits data with so small
residuals. Let us be not so faithful in suggestion that law of growth follows non-logistic
curve before checking through rigorous fitting procedure. In other words, "Did you fit
curve to your data? If yes, show me your fit and we can see together is it a new pattern
to grow or some other issue."
I was fascinated how many times forecasting with logistic S-curves leaded to reliable
forecasts. Four years ago, before studying literature, I could not imagine that dozen
thousands of S-curve fits represent reality so accurate way.
Fitting S-curve on data does not mean automatically successful forecast. For instance,
T.Modis published his book [39] in 1992. The second version of this book [13] was
published 10 years later with detail analysis of fault predictions made in first one. What
is interesting to learn, there were not so many faults. About 70% of presented forecasts
are correct. The book is full of logistic curves fitted to data.
In close look, the famous law of ideality growth [15] is a corollary of logistic S-curve
which represents law of natural growth.
In order to conclude the answer for the question I would like to propose, that it is
essential to apply a quantitative way to construct S-curves. As soon we draw our 'S-
curves' by hand, we do not present the trends, but our unverifiable hypothesis. It does
not matter how these hypotheses are genius, as soon they are unverifiable, they are not
reliable. As result they are not applicable.
Question:
1. Am I wrong or do you pose that TRIZ-publication in English can only decrease from
now on?
2. What is your opinion about the second26
Moore's law?
DK:
First question: In accordance with the results of fit S-curve to the available data, TRIZ-
publication in English for selected sources will not grow the same way in coming
future. Please look to the presented diagrams carefully. They are not so dramatic for
coming two years. We can speculate "why so?" It looks like, as usually, we have a
competition issue… In coming future we will verify together how the presented forecast
is correct.
Second question: One of the roles of each forecast (it was discussed in scope of tutorial
[40] in section "Why do we need to forecast?") is to increase awareness. Actually, the
predictive power of Moore's law is based on the fact that 'the more widely it became
accepted, the more it served as a goal for an entire industry.' There is a kind of self-
26
The notion 'second' was not understood exactly.
26
27. fulfilling prophecy phenomenon. Last years, many companies struggle to keep up with
Moore's law.
Initially the Moore's law was formulated as exponential growth. (Originally it was
suggested that the number of transistors that can be inexpensively placed on an
integrated circuit will be doubling annually.) Later in was corrected (towards 24 months
for doubling. Different sources refer to 18 months for doubling the number of
transistors.) Moore's law name is applied as a synonym to exponential growth especially
in computer technologies. Exponential growth is a part of logistic natural growth. If we
consider these facts it becomes clear, why this particular prediction became so famous.
According to the evolution of many generation of systems there is obvious, that system
cannot evolve endless exponentially. The same is occurring with integrated circuit. The
number of transistors approaches its ceiling, and trend changes from exponential one to
logistic curve.
In accordance with what was described many years ago [3, 4, 15], the rhythm of
changes from exponential growth to logistic one depends from how many resources will
be sacrificed to keep up with Moore's law. The mechanism was well described by G.
Altshuller in the book Creativity as Exact Science [15] and explained through many
examples in multiple TRIZ-publications.
"…The emphasis of logistic curves analysis becomes not building
or running the model, but interpreting it…" adapted from [25]
Contact:
* Dmitry KUCHARAVY
LICIA / LGECO, INSA Strasbourg
24, Bd de la Victoire, 67084 Strasbourg, France
tel: +33 (0)3 88 14 47 10; fax: +33 (0)3 88 14 47 99
E-mail: dmitry.kucharavy@insa-strasbourg.fr
Dmitry KUCHARAVY is a research engineer of INSA Strasbourg - Graduate School of
Science and Technology, France. He is a doctoral student at the University of Louis
Pasteur (Ecole Doctorale des Sciences pour l'Ingénieur). His research interests are in
Technology Forecasting of systems evolution, Theory of Advanced Thinking (OTSM-
TRIZ), and Problem solving knowledge management.
Roland DE GUIO is professor at the National Institute of Applied Science (INSA) of
Strasbourg, France. He worked 13 years in the area of application of operational
research and data analysis techniques to production flow analysis and design problems.
Since 2000, his main interest is developing holistic approaches, technologies and tools
that support innovative design in the frame of the research team in innovative design at
INSA.
27
28. REFERENCES
1. Christensen, C.M. The Innovator's Dilemma: the revolutionary national
bestseller that change the way we do business. (Harper Business Essentials, New
York, 2003), 287. ISBN 0-06-052199-6.
2. Christensen, C.M. Exploring the Limits of the Technology S-curve. Part I:
Component Technologies. Production and Operations Management, 1992, 1(4),
334-357.
3. Ayres, R.U. Technological Forecasting and Long-Range Planning. (McGraw-
Hill book Company, 1969), 237. ISBN 0070026637.
4. Martino, J.P. Technological Forecasting for Decision Making. (Mcgraw-Hill
(Tx), 1992), 462. ISBN 978-0070407770.
5. Kucharavy, D. TRIZ instruments for forecasting: past, present, and future. p. 80
(INSA Strasbourg - Graduate School of Science and Technology, Strasbourg,
2007). Tutorial at ETRIA TFC 2007, Nov. 6-8. Frankfurt, Germany.
6. Modis, T. A Scientific Approach to Managing Competition. The Industrial
Physicist, 2003, 9(1), 24-27.
7. Smitalova, K.; Sujan, S. A Mathematical Treatment of Dynamical Models in
Biological Science; Ellis Horwood: West Sussex, U.K., 1991; 183 pp.
8. Makridakis, S. and Hibon, M. The M3-Competition: results, conclusions and
implications. International Journal of Forecasting, 2000, 16, 451-476.
9. Meyer, P.S., Yung, J.W. and Ausubel, J.H. A Primer on Logistic Growth and
Substitution: The Mathematics of the Loglet Lab Software. Technological
Forecasting and Social Change, 1999, 61(3), 247-271.
10. Ayres, R.U. What have we learned? Technological Forecasting and Social
Change, 1999, 62(1-2), 9-12.
11. Modis, T. Strengths and weaknesses of S-curves. Technological Forecasting and
Social Change, 2007, 74(6), 866-872.
12. Young, P. Technological growth curves: A competition of forecasting models.
Technological Forecasting and Social Change, 1993, 44(4), 375-389.
13. Modis, T. Predictions - 10 Years Later. (Growth Dynamics, Geneva,
Switzerland, 2002), 335. ISBN 2-9700216-1-7.
14. Armstrong, J.S. Long Range Forecasting. From Crystal Ball to Computer. (John
Wiley & Sons, Inc., 1985), 689. ISBN 0-471-82360-0.
15. Altshuller, G.S. Creativity as an Exact Science. (Sovietskoe radio Publishing
House, Moscow, 1979), 184. (in English: Altshuller, G.S. Creativity as an Exact
Science: The Theory of the Solution of Inventive Problems. (Gordon and Breach
Science Publishers, 1984), 320. ISBN 0-677-21230-5.)
16. Fisher, J.C. and R.H. Pry, A simple substitution model of technological change.
Technological Forecasting and Social Change, 1971. 3(1): p. 75-78.
17. Kucharavy, D. and De Guio, R. Application of S-Shaped Curves. 7th ETRIA
TRIZ Future Conference (Kassel University Press GmbH, Kassel, Frankfurt,
Germany, 2007).
28
29. 18. Rogers, E.M. Diffusion of Innovations. (Free Press, New York, 2003), 512.
ISBN 0743222091.
19. Principles of Forecasting: A Handbook for Researchers and Practitioner.
(Kluwer Academic Publishers, Boston / Dordrecht / London, 2002), 849. ISBN
0-7923-7930-6.
20. Glenn, J.C. and Gordon, T.J., eds. Futures Research Methodology. (AC/UNU
Millennium Project, Washington, D.C. 20016-4055, 2003).
21. Porter, A.L. and Cunningham, S.W. Tech mining: exploiting new technologies
for competitive advantage. (John Willey & Sons Inc., Hoboken, New Jersey,
2005), 384. ISBN 0-471-47567-X.
22. Molitor, G.T.T. Molitor Forecasting Model: Key Dimensions for Plotting the
"Patterns of Change". Journal of Future Studies, 2003, 8(1), 61-72.
23. Mensch, G. Stalemate in Technology: Innovations Overcome the Depression
(Ballinger Pub Co, Cambridge, Massachusetts, 1978), 241. 088410611X.
24. Marchetti, C. Infrastructures for movement. Technological Forecasting and
Social Change, 1987, 32(4), 373-393.
25. Meyer, P.S., Yung, J.W. and Ausubel, J.H. A Primer on Logistic Growth and
Substitution: The Mathematics of the Loglet Lab Software. Technological
Forecasting and Social Change, 1999, 61(3), 247-271.
26. Altshuller, G.S. About forecasting of technical systems development. Seminars
materials p. 8 (Baku, 1975). In Russian
27. Gahide, S. Application of TRIZ to Technology Forecasting. Case Study: Yarn
Spinning Technology. (TRIZ Journal, www.triz-journal.com, 2000).
28. Gibson, N. The Determination of the Technological Maturity of Ultrasonic
Welding. (TRIZ Journal, www.triz-journal.com, 1999).
29. Slocum, M.S. Technology Maturity Using S-curve Descriptors. (TRIZ Journal,
www.triz-journal.com, 1999).
30. Stelzner, J., Palacios, C. and Swaton, T. TRIZ on Rapid Prototyping - a case
study for technology foresight. (TRIZ Journal, www.triz-journal.com, 2003).
31. Lovel, K., Seastrunk, C. and Clapp, T. The Application of TRIZ to Technology
Forecasting. A Case Study: Brassiere Strap Technology. (TRIZ Journal,
www.triz-journal.com, 2006).
32. Mitchell, I. and Mann, D. Overcoming Limiting Contradictions in a Continuous
Manufacturing Process. World Conference TRIZ Future 2002, pp. 139-152
(ETRIA, European TRIZ Association, Strasbourg, France, 2002).
33. Gautier, L., Marrony, M. and Kucharavy, D. Technological forecasting of Fuel
Cells for small stationary applications. p. 38 (European Institute for Energy
Research - EIfER, Karlsruhe, 2005). Report.
34. March, J.G. Exploration and Exploitation in Organizational Learning.
Organization Science, 1991, 2(1), 71-87.
35. Llerena, P. and Schenk, E. Technology Policy and A-Synchronic Technologies:
The Case of German High-Speed Trains. In Llerena, P. and Matt, M., eds.
29
30. Innovation Policy in a Knowledge Based Economy, pp. 115-134 (Springer,
Berlin, 2005).
36. Miller, G.A. The Magical Number Seven, Plus or Minus Two: Some Limits on
Our Capacity for Processing Information. The Psychological Review, 1956, 63,
81-97.
37. Cowan, R., Jonard, N. and Ozman, M. Knowledge dynamics in a network
industry. Technological Forecasting and Social Change, 2004, 71(5), 469-484.
38. Gaba, A. Risk assessment: Keep it simple. (INSEAD Knowledge 2007).
39. Modis, T. Predictions - Society's Telltale Signature Reveals the Past and
Forecasts the Future. (Simon & Schuster, New York, 1992), 300. ISBN
0-671-75917-5.
40. Kucharavy, D. TRIZ instruments for forecasting: past, present, and future. p. 80
(INSA Strasbourg - Graduate School of Science and Technology, Strasbourg,
2007). Manuscript.
30