Why does large number of data warehousing projects fail? How to avoid such failures? How to meet out user’s expectations and fulfil data analysis needs of business managers from data warehousing solutions? How to make data warehousing projects successful? These are some of the key questions before data warehouse research community in the present time. Literature shows that large numbers of data warehousing projects undertaken eventually result in a failure. In this paper, we have designed a framework named “Data Warehouse Development Standardization Framework” (DWDSF), to help data warehouse developer’s community in implementing effective data warehousing solutions. We have critically analysed literature to find out possible reasons of data warehouse project failure. Our framework has been designed to overcome such issues and enable implementation of successful data warehousing solutions. To verify usefulness of our framework, we have applied guidelines of DWDSF framework to design and implement data warehousing solution for National Rural Health Mission (NRHM) project which offers various health services throughout the country. The developed solution is giving results for all type of queries business managers want to run. We have shown results of some sample queries executed over the implemented data warehouse repository. All results are meeting out business manager’s query expectations.
Role of Operational System Design in Data Warehouse Implementation: Identifyi...iosrjce
Data warehouse designing process takes input from operational system of the organization. Quality
of data warehousing solution depends on design of operational system. Often, operational system
implementations of organizations have some limitations. Thus, we cannot proceed for data warehouse
designing so easily. In this paper, we have tried to investigate operational system of the organization for
identifying such limitations and determine role of operational system design in the process of data warehouse
design and implementation. We have worked out to find possible methods to handle such limitations and have
proposed techniques to get a quality data warehousing solution under such limitations. To make the work based
on live example, National Rural Health Mission (NRHM) Project has been taken. It is a national project of
health sector, managed by Indian Government across the country. The complex structure and high volume of
data makes it an ideal case for data warehouse implementation.
Designing a Framework to Standardize Data Warehouse Development Process for E...ijdms
Data warehousing solutions work as information base for large organizations to support their decision
making tasks. With the proven need of such solutions in current times, it is crucial to effectively design,
implement and utilize these solutions. Data warehouse (DW) implementation has been a challenge for the
organizations and the success rate of its implementation has been very low. To address these problems, we
have proposed a framework for developing effective data warehousing solutions. The framework is
primarily based on procedural aspect of data warehouse development and aims to standardize its process.
We first identified its components and then worked on them in depth to come up with the framework for
effective implementation of data warehousing projects. To verify effectiveness of the designed framework,
we worked on National Rural Health Mission (NRHM) project of Indian government and designed data
warehousing solution using the proposed framework.
KEYWORDS
Data warehousing, Framework Design, Dimensional
Data warehouse generally contains both types of data i.e. historical & current data from various data sources. Data warehouse in world of computing can be defined as system created for analysis and reporting of these both types of data. These analysis report is then used by an organization to make decisions which helps them in their growth. Construction of data warehouse appears to be simple, collection of data from data sources into one place (after extraction, transform and loading). But construction involves several issues such as inconsistent data, logic conflicts, user acceptance, cost, quality, security, stake holder’s contradictions, REST alignment etc. These issues need to be overcome otherwise will lead to unfortunate consequences affecting the organization growth. Proposed model tries to solve these issues such as REST alignment, stake holder’s contradiction etc. by involving experts of various domains such as technical, analytical, decision makers, management representatives etc. during initialization phase to better understand the requirements and mapping these requirements to data sources during design phase of data warehouse.
Data-Ed Online: Engineering Solutions to Data Quality ChallengesData Blueprint
This webinar originally aired on Tuesday, October 9th, 2012. It is part of Data Blueprint's ongoing webinar series on data management with Dr. Aiken.
Sign up for future sessions at http://www.datablueprint.com/webinar-schedule.
Abstract:
This presentation provides guidance to organizations considering or preparing for data quality initiatives. We will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality can be engineered provides a useful framework in which to develop an organizational approach. This in turn will allow organizations to more quickly identify data problems caused by structural issues versus practice-oriented defects. Participants will also learn the importance of practicing data quality engineering quantification.
Role of Operational System Design in Data Warehouse Implementation: Identifyi...iosrjce
Data warehouse designing process takes input from operational system of the organization. Quality
of data warehousing solution depends on design of operational system. Often, operational system
implementations of organizations have some limitations. Thus, we cannot proceed for data warehouse
designing so easily. In this paper, we have tried to investigate operational system of the organization for
identifying such limitations and determine role of operational system design in the process of data warehouse
design and implementation. We have worked out to find possible methods to handle such limitations and have
proposed techniques to get a quality data warehousing solution under such limitations. To make the work based
on live example, National Rural Health Mission (NRHM) Project has been taken. It is a national project of
health sector, managed by Indian Government across the country. The complex structure and high volume of
data makes it an ideal case for data warehouse implementation.
Designing a Framework to Standardize Data Warehouse Development Process for E...ijdms
Data warehousing solutions work as information base for large organizations to support their decision
making tasks. With the proven need of such solutions in current times, it is crucial to effectively design,
implement and utilize these solutions. Data warehouse (DW) implementation has been a challenge for the
organizations and the success rate of its implementation has been very low. To address these problems, we
have proposed a framework for developing effective data warehousing solutions. The framework is
primarily based on procedural aspect of data warehouse development and aims to standardize its process.
We first identified its components and then worked on them in depth to come up with the framework for
effective implementation of data warehousing projects. To verify effectiveness of the designed framework,
we worked on National Rural Health Mission (NRHM) project of Indian government and designed data
warehousing solution using the proposed framework.
KEYWORDS
Data warehousing, Framework Design, Dimensional
Data warehouse generally contains both types of data i.e. historical & current data from various data sources. Data warehouse in world of computing can be defined as system created for analysis and reporting of these both types of data. These analysis report is then used by an organization to make decisions which helps them in their growth. Construction of data warehouse appears to be simple, collection of data from data sources into one place (after extraction, transform and loading). But construction involves several issues such as inconsistent data, logic conflicts, user acceptance, cost, quality, security, stake holder’s contradictions, REST alignment etc. These issues need to be overcome otherwise will lead to unfortunate consequences affecting the organization growth. Proposed model tries to solve these issues such as REST alignment, stake holder’s contradiction etc. by involving experts of various domains such as technical, analytical, decision makers, management representatives etc. during initialization phase to better understand the requirements and mapping these requirements to data sources during design phase of data warehouse.
Data-Ed Online: Engineering Solutions to Data Quality ChallengesData Blueprint
This webinar originally aired on Tuesday, October 9th, 2012. It is part of Data Blueprint's ongoing webinar series on data management with Dr. Aiken.
Sign up for future sessions at http://www.datablueprint.com/webinar-schedule.
Abstract:
This presentation provides guidance to organizations considering or preparing for data quality initiatives. We will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality can be engineered provides a useful framework in which to develop an organizational approach. This in turn will allow organizations to more quickly identify data problems caused by structural issues versus practice-oriented defects. Participants will also learn the importance of practicing data quality engineering quantification.
The business dimensional life cycle. Summarized from the second chapter of 'The Data Warehouse Lifecyle Toolkit : Expert Methods for Designing, Developing, and Deploying Data Warehouses' by Ralph Kimball
A simplified approach for quality management in data warehouseIJDKP
Data warehousing is continuously gaining importance as organizations are realizing the benefits of
decision oriented data bases. However, the stumbling block to this rapid development is data quality issues
at various stages of data warehousing. Quality can be defined as a measure of excellence or a state free
from defects. Users appreciate quality products and available literature suggests that many organization`s
have significant data quality problems that have substantial social and economic impacts. A metadata
based quality system is introduced to manage quality of data in data warehouse. The approach is used to
analyze the quality of data warehouse system by checking the expected value of quality parameters with
that of actual values. The proposed approach is supported with a metadata framework that can store
additional information to analyze the quality parameters, whenever required.
A ROBUST APPROACH FOR DATA CLEANING USED BY DECISION TREEijcsa
Now a day’s every second trillion of bytes of data is being generated by enterprises especially in internet.To achieve the best decision for business profits, access to that data in a well-situated and interactive way is always a dream of business executives and managers. Data warehouse is the only viable solution that can bring the dream into veracity. The enhancement of future endeavours to make decisions depends on the availability of correct information that is based on quality of data underlying. The quality data can only be produced by cleaning data prior to loading into data warehouse since the data collected from different sources will be dirty. Once the data have been pre-processed and cleansed then it produces accurate results on applying the data mining query. Therefore the accuracy of data is vital for well-formed and reliable decision making. In this paper, we propose a framework which implements robust data quality to ensure consistent and correct loading of data into data warehouses which ensures accurate and reliable data analysis, data mining and knowledge discovery.
Leveraging Automated Data Validation to Reduce Software Development Timeline...Cognizant
Our enterprise solution for automating data validation - called dataTestPro - facilitates quality assurance (QA) by managing heterogeneous data testing, improving test scheduling, increasing data testing speed and reducing data -validation errors drastically.
Requirements engineering (RE), as a part of the project development life cycle, has increasingly been recognized as the key to ensure on-time, on-budget, and goal-based delivery of software projects. RE of big data projects is even more crucial because of the rapid growth of big data applications over the past few years. Data processing, being a part of big data RE, is an essential job in driving big data RE process successfully. Business can be overwhelmed by data and underwhelmed by the information so, data processing is very critical in big data projects. Employing traditional data processing techniques lacks the invention of useful information because of the main characteristics of big data, including high volume, velocity, and variety. Data processing can be benefited by process mining, and in turn, helps to increase the productivity of the big data projects. In this paper, the capability of process mining in big data RE to discover valuable insights and business values from event logs and processes of the systems has been highlighted. Also, the proposed big data requirements engineering framework, named REBD, helps software requirements engineers to eradicate many challenges of big data RE.
Data Profiling, Data Catalogs and Metadata HarmonisationAlan McSweeney
These notes discuss the related topics of Data Profiling, Data Catalogs and Metadata Harmonisation. It describes a detailed structure for data profiling activities. It identifies various open source and commercial tools and data profiling algorithms. Data profiling is a necessary pre-requisite activity in order to construct a data catalog. A data catalog makes an organisation’s data more discoverable. The data collected during data profiling forms the metadata contained in the data catalog. This assists with ensuring data quality. It is also a necessary activity for Master Data Management initiatives. These notes describe a metadata structure and provide details on metadata standards and sources.
IMPORTANCE OF PROCESS MINING FOR BIG DATA REQUIREMENTS ENGINEERINGijcsit
Requirements engineering (RE), as a part of the project development life cycle, has increasingly been recognized as the key to ensure on-time, on-budget, and goal-based delivery of software projects. RE of big data projects is even more crucial because of the rapid growth of big data applications over the past few years. Data processing, being a part of big data RE, is an essential job in driving big data RE process successfully. Business can be overwhelmed by data and underwhelmed by the information so, data processing is very critical in big data projects. Employing traditional data processing techniques lacks the invention of useful information because of the main characteristics of big data, including high volume, velocity, and variety. Data processing can be benefited by process mining, and in turn, helps to increase the productivity of the big data projects. In this paper, the capability of process mining in big data RE to discover valuable insights and business values from event logs and processes of the systems has been highlighted. Also, the proposed big data requirements engineering framework, named REBD, helps software requirements engineers to eradicate many challenges of big data RE.
Do you think data warehousing has failed in the market?? If yes, what are the possible causes?
What is the current status of data warehousing tools? What is required in future upcoming tools?
What is the difference between creation of data warehouse and implementation of MDDB?
The business dimensional life cycle. Summarized from the second chapter of 'The Data Warehouse Lifecyle Toolkit : Expert Methods for Designing, Developing, and Deploying Data Warehouses' by Ralph Kimball
A simplified approach for quality management in data warehouseIJDKP
Data warehousing is continuously gaining importance as organizations are realizing the benefits of
decision oriented data bases. However, the stumbling block to this rapid development is data quality issues
at various stages of data warehousing. Quality can be defined as a measure of excellence or a state free
from defects. Users appreciate quality products and available literature suggests that many organization`s
have significant data quality problems that have substantial social and economic impacts. A metadata
based quality system is introduced to manage quality of data in data warehouse. The approach is used to
analyze the quality of data warehouse system by checking the expected value of quality parameters with
that of actual values. The proposed approach is supported with a metadata framework that can store
additional information to analyze the quality parameters, whenever required.
A ROBUST APPROACH FOR DATA CLEANING USED BY DECISION TREEijcsa
Now a day’s every second trillion of bytes of data is being generated by enterprises especially in internet.To achieve the best decision for business profits, access to that data in a well-situated and interactive way is always a dream of business executives and managers. Data warehouse is the only viable solution that can bring the dream into veracity. The enhancement of future endeavours to make decisions depends on the availability of correct information that is based on quality of data underlying. The quality data can only be produced by cleaning data prior to loading into data warehouse since the data collected from different sources will be dirty. Once the data have been pre-processed and cleansed then it produces accurate results on applying the data mining query. Therefore the accuracy of data is vital for well-formed and reliable decision making. In this paper, we propose a framework which implements robust data quality to ensure consistent and correct loading of data into data warehouses which ensures accurate and reliable data analysis, data mining and knowledge discovery.
Leveraging Automated Data Validation to Reduce Software Development Timeline...Cognizant
Our enterprise solution for automating data validation - called dataTestPro - facilitates quality assurance (QA) by managing heterogeneous data testing, improving test scheduling, increasing data testing speed and reducing data -validation errors drastically.
Requirements engineering (RE), as a part of the project development life cycle, has increasingly been recognized as the key to ensure on-time, on-budget, and goal-based delivery of software projects. RE of big data projects is even more crucial because of the rapid growth of big data applications over the past few years. Data processing, being a part of big data RE, is an essential job in driving big data RE process successfully. Business can be overwhelmed by data and underwhelmed by the information so, data processing is very critical in big data projects. Employing traditional data processing techniques lacks the invention of useful information because of the main characteristics of big data, including high volume, velocity, and variety. Data processing can be benefited by process mining, and in turn, helps to increase the productivity of the big data projects. In this paper, the capability of process mining in big data RE to discover valuable insights and business values from event logs and processes of the systems has been highlighted. Also, the proposed big data requirements engineering framework, named REBD, helps software requirements engineers to eradicate many challenges of big data RE.
Data Profiling, Data Catalogs and Metadata HarmonisationAlan McSweeney
These notes discuss the related topics of Data Profiling, Data Catalogs and Metadata Harmonisation. It describes a detailed structure for data profiling activities. It identifies various open source and commercial tools and data profiling algorithms. Data profiling is a necessary pre-requisite activity in order to construct a data catalog. A data catalog makes an organisation’s data more discoverable. The data collected during data profiling forms the metadata contained in the data catalog. This assists with ensuring data quality. It is also a necessary activity for Master Data Management initiatives. These notes describe a metadata structure and provide details on metadata standards and sources.
IMPORTANCE OF PROCESS MINING FOR BIG DATA REQUIREMENTS ENGINEERINGijcsit
Requirements engineering (RE), as a part of the project development life cycle, has increasingly been recognized as the key to ensure on-time, on-budget, and goal-based delivery of software projects. RE of big data projects is even more crucial because of the rapid growth of big data applications over the past few years. Data processing, being a part of big data RE, is an essential job in driving big data RE process successfully. Business can be overwhelmed by data and underwhelmed by the information so, data processing is very critical in big data projects. Employing traditional data processing techniques lacks the invention of useful information because of the main characteristics of big data, including high volume, velocity, and variety. Data processing can be benefited by process mining, and in turn, helps to increase the productivity of the big data projects. In this paper, the capability of process mining in big data RE to discover valuable insights and business values from event logs and processes of the systems has been highlighted. Also, the proposed big data requirements engineering framework, named REBD, helps software requirements engineers to eradicate many challenges of big data RE.
Do you think data warehousing has failed in the market?? If yes, what are the possible causes?
What is the current status of data warehousing tools? What is required in future upcoming tools?
What is the difference between creation of data warehouse and implementation of MDDB?
Data Mining: A prediction Technique for the workers in the PR Department of O...ijcseit
This paper presents the method of mining the data and which contains the information about the large
information about the PR (Panchayat Raj Department) of Orissa .we have focused some of the techniques
,approaches and different methodology’s of the demand forecasting .Every organizations are operated in
different places of the country. Each place of operation may generate a huge amount of data. In an
organization, worker prediction is the difficult task of the manager. It is the complex process not only
because its nature of feature prediction but also various approaches methodologies always makes user
confused. This paper aims to deal with the problem selection process.In this paper we have used some of
the approaches from literatures are been introduced and analyzed to find its suitable organization and
situation, Based on this we have designed with automatic selection function to help users make a
prejudgment The information about each approach will be showed to users with examples to help
understanding. This system also provides calculation function to help users work out a predication result.
Generally the new developed system has a more comprehensive functions compared with existing ones .It
aims to improve the accuracy of demand forecasting by implementing the forecasting algorithm. While it is
still a decision support system with no ability of make the final judgment. .The data warehouse is used in
the significant business value by improving the effectiveness of managerial decision-making. In an
uncertain and highly competitive business environment, the value of strategic information systems such as
these are easily recognized however in today’s business environment, efficiency or speed is not the only
key for competitiveness. This type of huge amount of data’s are available in the form of tera- to peta-bytes
which has drastically changed in the areas of science and engineering. To analyze, manage and make a
decision of such type of huge amount of data we need techniques called the data mining which will
transforming in many fields. We have implemented the algorithms in JAVA technology. This paper provides
the prediction algorithm (Linear Regression, result which will helpful in the further research.
Suggest an intelligent framework for building business process management [ p...ijseajournal
As companies enter into the digital world, information technology is playing a major role in bringing
process improvements to the forefront of business management. In the recent decades, many organizations
have struggled to redesign and improve their business processes to reduce their total cost. The main
contribution of this research study is to propose an intelligent framework that possesses the ability to
employ a database of best practices, business standards, and business activity history in order to permit the
manager to analyze and improve the design of the business processes.
In addition, the other objective of this research is to build a business process or workflow directly from its
process design logic in order to enable rapid process development and deployment. This procedure
requires some technical improvements of the business design, as it is mainly based on building the business
process using Microsoft Office Visio, which communicates the defined business process to the business
process management engine.
Creating a Successful DataOps Framework for Your Business.pdfEnov8
As data is universally important and has a major role in decision-making and other business operations, a strong data-driven culture has become extremely important for business organizations.
This calls for a successful and efficient DataOps framework. Let us explore more about this emerging methodology.
Running head Database and Data Warehousing design1Database and.docxhealdkathaleen
Running head: Database and Data Warehousing design 1
Database and Data Warehousing Design 3
Database and Data Warehousing Design
Thien Thai
CIS599
Professor Wade M. Poole
Strayer University
Feb 20, 2020
Database and Data Warehousing Design
Introduction
Technology has highly revolutionized the world of business –hence presenting more challenges and opportunities for businesses. Companies which fail to embrace and incorporate technology in their operations risks being edged out of the market due to stiff competition witnessed in the market today. On the flipside, cloud-based technology allows businesses to “easily retrieve and store valuable data about their customers, products, and employees.” Data is an important component that help to support core business decisions. In today’s highly competitive and constantly evolving business world, embracing cloud-based technology business managers an opportunity to make informed and result-oriented decisions regarding day-to-day organizational operations (Dimitriu & Matei, 2015).
Notably, business growth and competitiveness depends on its ability to transform data into information. Data warehousing and adoption of relational databases are some of cloud-based technologies which have positively impacted on businesses. The two technologies have had a strategic value to companies –helping them to have the extra edge over their competitors. Both data warehousing and relational databases help businesses to “take smart decisions in a smarter manner.” However, failure to adopt these cloud-based technologies has hindered business executives’ ability to make experienced-based and fact-based decisions which are vital to business survival. Both “databases and data warehouses are relational data systems” which serve different and equally crucial roles within an organization. For instance, data warehousing helps to support management decisions while relational databases help to perform ongoing business transactions in real-time. Basically, embracing cloud-based technologies within the organization will help to give the company a competitive advantage in the market. However, the adoption and maintenance of such technologies require full support and endorsement of the business management. Organizational management must understand the feasibility, functionality, and the importance of embracing such technologies. Movement towards relational databases and data warehousing requires a lot of funding –hence the need to convince the management to support and fund them. This paper seeks to explore the concepts of data warehousing, relational databases, their importance to the business, as whey as their design.
“Importance of Data Warehousing and Relational Databases”
Today, technology has changed the market landscape. Business are striving to adopt cloud-based technology in order to improve efficiency in business functions –among them analytical queries as well as transactional operations. Both relational databases a ...
Running head Database and Data Warehousing design1Database and.docxtodd271
Running head: Database and Data Warehousing design 1
Database and Data Warehousing Design 3
Database and Data Warehousing Design
Thien Thai
CIS599
Professor Wade M. Poole
Strayer University
Feb 20, 2020
Database and Data Warehousing Design
Introduction
Technology has highly revolutionized the world of business –hence presenting more challenges and opportunities for businesses. Companies which fail to embrace and incorporate technology in their operations risks being edged out of the market due to stiff competition witnessed in the market today. On the flipside, cloud-based technology allows businesses to “easily retrieve and store valuable data about their customers, products, and employees.” Data is an important component that help to support core business decisions. In today’s highly competitive and constantly evolving business world, embracing cloud-based technology business managers an opportunity to make informed and result-oriented decisions regarding day-to-day organizational operations (Dimitriu & Matei, 2015).
Notably, business growth and competitiveness depends on its ability to transform data into information. Data warehousing and adoption of relational databases are some of cloud-based technologies which have positively impacted on businesses. The two technologies have had a strategic value to companies –helping them to have the extra edge over their competitors. Both data warehousing and relational databases help businesses to “take smart decisions in a smarter manner.” However, failure to adopt these cloud-based technologies has hindered business executives’ ability to make experienced-based and fact-based decisions which are vital to business survival. Both “databases and data warehouses are relational data systems” which serve different and equally crucial roles within an organization. For instance, data warehousing helps to support management decisions while relational databases help to perform ongoing business transactions in real-time. Basically, embracing cloud-based technologies within the organization will help to give the company a competitive advantage in the market. However, the adoption and maintenance of such technologies require full support and endorsement of the business management. Organizational management must understand the feasibility, functionality, and the importance of embracing such technologies. Movement towards relational databases and data warehousing requires a lot of funding –hence the need to convince the management to support and fund them. This paper seeks to explore the concepts of data warehousing, relational databases, their importance to the business, as whey as their design.
“Importance of Data Warehousing and Relational Databases”
Today, technology has changed the market landscape. Business are striving to adopt cloud-based technology in order to improve efficiency in business functions –among them analytical queries as well as transactional operations. Both relational databases a.
Successfully supporting managerial decision-making is critically dep.pdfanushasarees
Successfully supporting managerial decision-making is critically dependent upon the availability
of integrated, high quality information organized and presented in a timely and easily understood
manner. Data warehouses have emerged to meet this need. They serve as an integrated repository
for internal and external data—intelligence critical to understanding and evaluating the business
within its environmental context. With the addition of models, analytic tools, and user interfaces,
they have the potential to provide actionable information resources—business intelligence that
supports effective problem and opportunity identification, critical decision-making, and strategy
formulation, implementation, and evaluation. Four themes frame our analysis: integration,
implementation, intelligence, and innovation.
1:four major categories of business environment factors is
INTEGRATION,IMPLEMENTATION,INTELLIGENCE AND INNOVATION.
Organizations use data warehousing to support strategic and mission-critical applications. Data
deposited into the data warehouse must be transformed into information and knowledge and
appropriately disseminated to decision-makers within the organization and to critical partners in
various capacities within the organizational value chain. Crucial problems that must be addressed
in this area are: the modes of dissemination of information to the end user; the development,
selection, and implementation of appropriate models, analytic tools, and data mining tools; the
privacy and security of data; system performance; and adequate levels of training and support.
The human–computer interface is of paramount importance in the data warehouse environment
and the primary determinant of success from the end-user perspective. In order to support
analysis and reporting tasks, the data warehouse must have high quality data and make these data
accessible through intuitive interface technologies. Data warehouse browsing tools provide star-
schema query-like access through a flexible menu-based interface, with pull-down menus
representing important dimensions. These types of tools are easy to use and support some ad-hoc
exploration, but are usually controlled through an administrative layer that determines the data
available to endusers. In developing a flexible interface, there is a tradeoff between the ability to
express ad-hoc queries and the ease-of-use that results from pre-defined constructs implemented
by data warehouse designers and administrators. Of course, SQL can provide an ad-hoc query
facility, but its use requires some care in the data warehouse environment where the combination
of very large tables and ill-formed user queries can produce some truly awful performance and
potentially erroneous results. Casual users may not have sufficient understanding of SQL or of
the database schema to effectively use such an interface. Typically, only trained power users
(e.g., DBAs, application developers) are permitted to write SQL queries on .
Making the right decision at the right time requires that the data should be available at internet time not
after the event when it is too late to do anything. So the need for a web portal has been increased in order
to support decision making in the different organizational levels at internet time by providing features
such as, customization, personalization, support for collaboration and notification that support managers
in making the right decision at internet time. All this can be achieved by applying the (BI) tools on “One
version” data store that contains data from legacy ERP system after applying the ETL tools over it. Also
the Data marts will be used in order to shorten the response time of queries generated by the users.
Turning your Excel Business Process Workflows into an Automated Business Inte...OAUGNJ
Many organizations have evolved key internal business processes built on top of Microsoft Excel. These cross-functional workflows involve several organizational units responsible for collecting business system transactions, modifying this raw data, consolidating, transforming, pivoting and preparing data into a published set of Reports & Graphs – all in MS Excel. Such workflows are a burden to organizations – not repeatable, costly, time-consuming, inflexible and hard to scale, and evolve to become more complex over time. Business critical processes such as financial analysis, operational analysis and revenue analysis are often supported this way. Attempting to replace such systems can be quite daunting and a barrier to replace. The goal of this session is to present an easy to understand methodology and use cases to demonstrate how to move from an operational workflow in Excel to truly automated Business Intelligence.
Requirements engineering (RE), as a part of the project development life cycle, has increasingly been
recognized as the key to ensure on-time, on-budget, and goal-based delivery of software projects. RE of big
data projects is even more crucial because of the rapid growth of big data applications over the past few
years. Data processing, being a part of big data RE, is an essential job in driving big data RE process
successfully. Business can be overwhelmed by data and underwhelmed by the information so, data
processing is very critical in big data projects. Employing traditional data processing techniques lacks the
invention of useful information because of the main characteristics of big data, including high volume,
velocity, and variety. Data processing can be benefited by process mining, and in turn, helps to increase
the productivity of the big data projects. In this paper, the capability of process mining in big data RE to
discover valuable insights and business values from event logs and processes of the systems has been
highlighted. Also, the proposed big data requirements engineering framework, named REBD, helps
software requirements engineers to eradicate many challenges of big data RE
Similar to Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data Warehouse Failure (20)
Land Cover maps supply information about the physical material at the surface of the Earth (i.e. grass, trees, bare ground, asphalt, water, etc.). Usually they are 2D representations so to present variability of land covers about latitude and longitude or other type of earth coordinates. Possibility to link this variability to the terrain elevation is very useful because it permits to investigate probable correlations between the type of physical material at the surface and the relief. This paper is aimed to describe the approach to be followed to obtain 3D visualizations of land cover maps in GIS (Geographic Information System) environment. Particularly Corine Land Cover vector files concerning Campania Region (Italy) are considered: transformed raster files are overlapped to DEM (Digital Elevation Model) with adequate resolution and 3D visualizations of them are obtained using GIS tool. The resulting models are discussed in terms of their possible use to support scientific studies on Campania Land Cover.
Comparison between Cisco ACI and VMWARE NSXIOSRjournaljce
Software-Defined Networking(SDN) allows you to have a logical image of the components in the data center, also you could arrange the components logically and use them according to the software application needs. This paper gives an overview about the architectural features of Cisco’s Application Centric Infrastructure (ACI) and Vmware’s NSX and also compares both the architectures and their benefit
Student’s Skills Evaluation Techniques using Data Mining.IOSRjournaljce
Technological Advancement is the mainstay of the Indian economy. Software Development is a primary factor which offers many sources of employment across the state. The major role of Software Company is to provide qualitized product by Managerial Decisions. To improve the Software Quality and Maintenance, Organization concentrates the Programmer’s ability and restricts the Placement with various kinds of audiences such as Written test, Personal Interview, Technical Round, Group Discussion etc. Accordingly, Programmer’s has to be formed. So Higher Educational institutes have to amend the student’s Performance as their expectations. The student’s Programming Skill is analyzed with the help of Decision making and Knowledge Discovery in Data Mining Techniques. In this example, the actual data are collected from Private College, which holds data about academic performance for pupils. This report suggests a method to judge their accomplishments through the utilization of data mining methods and specifies the means to identify their skills and assist them to improve their Knowledge by predicting Training Programs.
Data mining is a process to extract information from a huge amount of data and transform it into an
understandable structure. Data mining provides the number of tasks to extract data from large databases such
as Classification, Clustering, Regression, Association rule mining. This paper provides the concept of
Classification. Classification is an important data mining technique based on machine learning which is used to
classify the each item on the bases of features of the item with respect to the predefined set of classes or groups.
This paper summarises various techniques that are implemented for the classification such as k-NN, Decision
Tree, Naïve Bayes, SVM, ANN and RF. The techniques are analyzed and compared on the basis of their
advantages and disadvantages
Analyzing the Difference of Cluster, Grid, Utility & Cloud ComputingIOSRjournaljce
: Virtualization and cloud computing is creating a fundamental change in computer architecture,
software and tools development, in the way we store, distribute and consume information. In the recent era of
autonomic computing it comes the importance and need of basic concepts of having and sharing various
hardware and software and other resources & applications that can manage themself with high level of human
guidance. Virtualization or Autonomic computing is not a new to the world, but it developed rapidly with Cloud
computing. In this paper there give an overview of various types of computing. There will be discussion on
Cluster, Grid computing, Utility & Cloud Computing. Analysis architecture, differences between them,
characteristics , its working, advantages and disadvantages
: Cloud processing is turning into an inexorably mainstream endeavor demonstrate in which figuring assets are made accessible on-request to the client as required. The one of a kind incentivized offer of distributed computing makes new chances to adjust IT and business objectives. Distributed computing utilizes the web advancements for conveyance of IT-Enabled abilities 'as an administration' to any required clients i.e. through distributed computing we can get to anything that we need from anyplace to any PC without agonizing over anything like about their stockpiling, cost, administration etc. In this paper, I give a far-reaching study on the inspiration variables of receiving distributed computing, audit the few cloud sending and administration models. It additionally investigates certain advantages of distributed computing over customary IT benefit environment-including versatility, adaptability, decreased capital and higher asset usage are considered as appropriation explanations behind distributed computing environment. I additionally incorporate security, protection, and web reliance and accessibility as shirking issues. The later incorporates vertical versatility as specialized test in cloud environment.
An Experimental Study of Diabetes Disease Prediction System Using Classificat...IOSRjournaljce
Data mining means to the process of collecting, searching through, and analyzing a large amount of data in a database. Classification in one of the well-known data mining techniques for analyzing the performance of Naive Bayes, Random Forest, and Naïve Bayes tree (NB-Tree) classifier during the classification to improve precision, recall, f-measure, and accuracy. These three algorithms, of Naive Bayes, Random Forest, and NB-Tree are useful and efficient, has been tested in the medical dataset for diabetes disease and solving classification problem in data mining. In this paper, we compare the three different algorithms, and results indicate the Naive Bayes algorithms are able to achieve high accuracy rate along with minimum error rate when compared to other algorithms.
Candidate Ranking and Evaluation System based on Digital FootprintsIOSRjournaljce
Digital resume provides insights about a candidate to the organization. This paper proposes a system where digital resumes of candidates are generated by extracting data from social networking sites like Facebook, Twitter and LinkedIn. Data which is relevant to recruitment is obtained from unstructured data using Data Mining algorithms. Candidates are evaluated based on their digital resumes and ranked accordingly. Ranking is done based on the requirements specified by an organization for a key position. The key aspects of this paper are a) Specification and design of system. b) Generation of digital Resume. c) Ranking of candidates. According to the ranking provided by this system, Recruiters can shortlist candidates for interviews. Thus, it revolutionizes the traditional recruitment process.
Multi Class Cervical Cancer Classification by using ERSTCM, EMSD & CFE method...IOSRjournaljce
Cervical cancer is the highest rate of incidence after breast cancer, gastric cancer, colorectal cancer, thyroid cancer among all malignant that occurs to females ; also it is the most prevalent cancer among female genital cancers. Manual cervical cancer diagnosis methods are costly and sometimes result inaccurate diagnosis caused by human error but machine assisted classification system can reduce financial costs and increase screening accuracy. In this research article, we have developed multi class cervical classification system by using Pap Smear Images according to the WHO descriptive Classification of Cervical Histology. Then, this system classifies the cell of the Pap Smear image into anyone of five types of the classes of normal cell, mild dysplasia, moderate dysplasia, severe dysplasia and carcinoma in situ (CIS) by using individual and Combining individual feature extraction method with the classification technique. In this paper three Feature Extraction methods were used: From that three, two were individual feature extraction method namely Enriched Rough Set Texton Co-Occurrence Matrix (ERSTCM) and Enriched Micro Structure Descriptor (EMSD) and the remained one was combining individual feature extraction method namely concatenated feature extraction method (CFE). The CFE method represents all the individual feature extraction methods of ERSTCM & EMSD features are combining together to one feature to assess their joint performance. Then these three feature extraction methods are tested over Fuzzy Logic based Hybrid Kernel Support Vector Machine (FL-HKSVM) Classifier. This Examination was conducted over a set of single cervical cell based pap smear images. The dataset contains five classes of images, with a total of 952 images. The distribution of number of images per class is not uniform. Then the performance was evaluated in both the individual and combining individual feature extraction method with the classification techniques by using the statistical parameters of sensitivity, specificity & accuracy. Hence the resultant values of the statistical parameters described in individual feature extraction method with the classification technique, proposed EMSD+FLHKSVM Classifier had given the better results than the other ERSTCM+FLHKSVM Classifier and combining individual feature extraction method with the classification technique described, proposed CFE+FLHKSVM Classifier had given the better results than other EMSD+FLHKSVM & ERSTCM+FLHKSVM classifiers.
The Systematic Methodology for Accurate Test Packet Generation and Fault Loca...IOSRjournaljce
As we probably aware now a days networks are broadly dispersed so administrators relies on upon different devices, for example, pings and follow course to troubleshoot the issue in network. So we proposed a robotized and orderly approach for testing and troubleshooting network called "Automatic Test Packet Generation"(ATPG). ATPG first peruses switch arrangement and produces a gadget free model. The model is utilized to produce least number of test packets to cover each connection in a network and every control in network. ATPG is equipped for researching both useful and execution issues. Test packets are sent at customary interims and separate strategy is utilized to confine flaws. The working of few disconnected devices which automatically create test packets are additionally given, yet ATPG goes past the prior work in static (checking liveness and fault localization).
The Use of K-NN and Bees Algorithm for Big Data Intrusion Detection SystemIOSRjournaljce
Big data problem in intrusion detection system is mainly due to the large volume of the data. The dimension of the original data is 41. Some of the feature of original data are unnecessary. In this process, the volume of data has expanded into hundreds and thousands of gigabytes(GB) of information. The dimension span of data and volume can be reduced and the system is enhanced by using K-NN and BA. The reduction ratio of KDD datasets and processing speed is very slow so the data has been reduced for extracting features by Bees Algorithm (AB) and use K-nearest neighbors as classification (KNN). So, the KDD99 datasets applied in the experiments with significant features. The results have gave higher detection and accuracy rate as well as reduced false positive rate. Keywords: Big Data; Intru
Study and analysis of E-Governance Information Security (InfoSec) in Indian C...IOSRjournaljce
The purpose of the study is to explore and find a research gap in E-Governance Information Security (InfoSec) domain in Indian Context. The study identifies the research gap in E-Governance InfoSec domain and substantiates given research gap with relevant literature review. The study outcomes clearly depict the requirement of research in the field of InfoSec in e-governance domain in a country like India.
The purpose of the study is to explore web security status of the E-Governance websites and web applications. The central theme of the paper is to study and analyze the security vulnerabilities in the technologies utilized for E-Governance website and web application development. The study was conducted in the State of Gujarat,India. The data related to web development technologies, vulnerabilities affecting those technologies, vulnerability severity, and vulnerability type were gathered from 26 E-Governance website/Web application for detail analysis. The outcome of the study depicts the relationship between technology vis-a-vis vulnerability type and vulnerability severity.
Exploring 3D-Virtual Learning Environments with Adaptive RepetitionsIOSRjournaljce
: In spatial tasks, the use of cognitive aids reduce mental load and therefore being appealing to trainers and trainees. However, these aids can act as shortcuts and prevents the trainees from active exploration which is necessary to perform the task independently in non-supervised environment. In this paper we used adaptive repetition as control strategy to explore the 3D- Virtual Learning environments. The proposed approach enables the trainee to get the benefits of cognitive support while at the same time he is actively involved in the learning process. Experimental results show the effectiveness of the proposed approach
Human Face Detection Systemin ANew AlgorithmIOSRjournaljce
Trying to detecting a face from any photo is big problem and got these days a focusing because of it importance, in face recognition system,face detection in one of the basic components. A lot of troubles are there to be solved in order to create a successful face detection algorithm. The skin of face has its properties in color domain and also a texture which may help in the algorithm for detecting faces because of its ability to find skins from photo. Here we are going to create a new algorithm for human face detection depending on skin color tone specially YCbCr color tone as an approach to slice the photo into parts. In addition, Gray level has been used to detect the area which contains a skin, after that anotherlevel used to erase the area that does not contain skin. The system which proposed applied on many photos and passed with great accuracy of detecting faces and it has a good efficient especially to separate the area that does not contain skin or face from the area which contain face and skin. It has been agreed and approved that the accuracy of the proposed system is 98% in human face detection.
Value Based Decision Control: Preferences Portfolio Allocation, Winer and Col...IOSRjournaljce
The paper presents an innovative approach to mathematical modeling of complex systems „humandynamical process”. The approach is based on the theory of measurement and utility theory and permits inclusion of human preferences in the objective function. The objective utility function is constructed by recurrent stochastic procedure which represents machine learning based on the human preferences. The approach is demonstrated by two case studies, portfolio allocation with Wiener process and portfolio allocation in the case of financial process with colored noise. The presented formulations could serve as foundation of development of decision support tools for design of management/control. This value-oriented modeling leads to the development of preferences-based decision support in machine learning environment and control/management value based design.
Assessment of the Approaches Used in Indigenous Software Products Development...IOSRjournaljce
Acceptability of indigenous software is always low in most of the African countries especially in Nigeria. This paper then study and presents the results on the assessment of the approaches used in indigenous software development products. The study involved ICT firms that specialized in software development and educational institutions who were part of major stakeholders as well as users of software packages. The primary tool for data collection was questionnaire, which was used to elicit information from software developers on the various approaches adopted in their operations. This was also complemented with information from secondary sources. The identified approaches were measured on a five-point Likert scale rating of 5 to 1 to determine their relative strength index (RSI) in the factors. The result revealed the various approaches adopted for software development had significant difference of chi (45)1699.06 at p≤ 0.001 with spiral (6.02), agile (5.86), prototyping (5.67), object oriented (5.48), rotational unified process (5.32), computer and incremental case (4.50), waterfall (3.66) and integrated (1.98) were the commonly adopted approaches used for software development. Similarly, the approaches adopted by software development firms were correlated and returned a significant difference of (Z = 1699.06, p≤ 0.001). The result implies that these approaches had a great impact on the domestic use of software products and perhaps is the most important driver of software industry growth for emerging technologies.
Secure Data Sharing and Search in Cloud Based Data Using Authoritywise Dynami...IOSRjournaljce
The Data sharing is an important functionality in cloud storage. We describe new public key crypto systems which produce constant-size cipher texts such that efficient delegation of decryption rights for any set of cipher texts are possible. The novelty is that one can aggregate any set of secret keys and make them as compact as a single key, but encompassing the power of all the keys being aggregated. Ensuring the security of cloud computing is second major factor and dealing with because of service availability failure the single cloud providers demonstrated less famous failure and possibility malicious insiders in the single cloud. A movement towards Multi-Clouds, In other words ”Inter-Clouds” or ”Cloud-Of-Clouds” as emerged recently. This works aim to reduce security risk and better flexibility and efficiency to the user. Multi-cloud environment has ability to reduce the security risks as well as it can ensure the security and reliability.
Panorama Technique for 3D Animation movie, Design and EvaluatingIOSRjournaljce
This paper presents an applied approach for Panorama 3D movies enhanced with visual sound effects. The case study that considered is IIPS@UOIT. Many selected S/W have been used to introduce the 3D Movie. 3D Animation is a modern technology in the field of the world of filmmaking and is considered the core of multimedia, where the vast majority of movies such as Hollywood movies that we see today, it was using 3D technology. Where this technique is used in all the magazines, such as medical experiments, engineering, astronomy, planets and stars, to prove scientific theories, history, geography, etc., where they are building models or scenes or characters simulates reality and the movement of the viewer to the heart of the event. A three-dimensional film was made to (IIPS @ UOITC) to give it a future vision and published in the global sites such as YouTube, Facebook and Google earth. By using many specialized 3d software and cinematic tricks, with a focusing on movement, characters, lighting, cameras and final render.
Density Driven Image Coding for Tumor Detection in mri ImageIOSRjournaljce
The significant of multi spectral band resolution is explored towards selection of feature coefficients based on its energy density. Toward the feature representiaon in transformed domain, multi wavelet transformations were used for finer spectral representation. However, due to a large feature count these features are not optimal under low resource computing system. In the recognition units, running with low resources a new coding approach of feature selection, considering the band spectral density is developed. The effective selection of feature element, based on its spectral density achieve two objective of pattern recognition, the feature coefficient representiaon is minimized, hence leading to lower resource requirement, and dominant feature representation, resulting in higher retrieval performance.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data Warehouse Failure
1. IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 19, Issue 1, Ver. II (Jan.-Feb. 2017), PP 29-38
www.iosrjournals.org
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 29 | Page
Data Warehouse Development Standardization Framework
(DWDSF): A Way to Handle Data Warehouse Failure
Deepak Asrani1
, Dr. Renu Jain2
, Dr. Usha Saxena3
1(Department of Computer Science Engineering, Teerthanker Mahaveer University, Moradabad, U.P (India))
2(Department of Computer Science Engineering, University Institute of Engg. & Tech, Kanpur, U.P (India))
3(State Institute of Health and Family Welfare, Indira Nagar, Lucknow, U.P (India))
Abstract: Why does large number of data warehousing projects fail? How to avoid such failures? How to meet
out user’s expectations and fulfil data analysis needs of business managers from data warehousing solutions?
How to make data warehousing projects successful? These are some of the key questions before data warehouse
research community in the present time. Literature shows that large numbers of data warehousing projects
undertaken eventually result in a failure. In this paper, we have designed a framework named “Data
Warehouse Development Standardization Framework” (DWDSF), to help data warehouse developer’s
community in implementing effective data warehousing solutions. We have critically analysed literature to find
out possible reasons of data warehouse project failure. Our framework has been designed to overcome such
issues and enable implementation of successful data warehousing solutions. To verify usefulness of our
framework, we have applied guidelines of DWDSF framework to design and implement data warehousing
solution for National Rural Health Mission (NRHM) project which offers various health services throughout the
country. The developed solution is giving results for all type of queries business managers want to run. We have
shown results of some sample queries executed over the implemented data warehouse repository. All results are
meeting out business manager’s query expectations.
Keywords: Data warehouse failure, Data warehouse framework, Data warehousing practices, Operational
System limitations
I. Introduction
Any business organization having quality operational system implemented is usually able to generate
useful MIS reports to visualize and manage its business processes and operations smoothly. But with the help of
operational system, business managers of the organization cannot generate reports useful for long term analysis
of their business processes and strategic decision making. Operational system reports do not have features to
produce deep insights of business operations to support strategic decision making. To meet out such
requirements one needs to have a quality data warehousing solution. From literature it is seen that major focus
of data warehousing research has been on improving physical implementation and run time query execution
performance. Limited attention is paid on issues related to data warehousing practices. Data warehousing
solution developers can ensure effectiveness in the developed solutions only if they follow a systematic
approach of data warehouse development procedure based on some well defined and tested framework that not
only offers specific guidelines about steps to be followed for data warehouse design and implementation but
also suggests the approach to be used during each step. We have designed a framework named “Data
Warehouse Development Standardization Framework” (DWDSF), to help data warehouse developer’s
community in implementing effective data warehousing solutions. We have experimentally verified its
usefulness. For dealing with issues of data warehouse failure, we first need to define and understand clearly
what actually is meaning of data warehouse failure. By data warehouse failure we mean not meeting out
specified goals and objectives of the organization from the developed solution. Research literature on data
warehouse failure identifies types of failures broadly categorized in two types. i. Failures due to managerial
aspects and ii.Failures due to technical aspects. Table 1 lists out some of the managerial aspects of data
warehouse failure and Table 2 lists out some of the technical aspects of data warehouse failure.
Table 1: List of managerial aspects of data warehouse failure
2. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 30 | Page
Table 2: List of technical aspects of data warehouse failure
If we closely analyze reasons of data warehousing solution’s failure, we find that most of the reasons
cited are related to issues of project development process (procedural issues). It is because efficient tools and
technologies are available at our disposal. We only need to rightly identify and follow a systematic approach of
data warehouse development procedure to ensure successful implementation of developed solution. We also
need to ensure utilization of available tools and technologies correctly. Planned development procedure is one
which is based on proper documentation, is well communicated, considers specific list of user expectations, and
is well estimated in terms of time and expenditure.Application development refers to computerization of
functioning of an organization or a system. Functioning is routine business activity which is almost always clear
and can be easily defined by the users. It is therefore easy to ensure its successful implementation. On the other
hand data warehousing project implementation refers to developing a platform of data repository to serve
business managers with variety of complex business queries for in depth analysis of all business subjects. In the
case of data warehouse implementation, requirements are generally not well defined by the users and therefore
needs to be determined carefully to succeed in implementation. Application development and data warehouse
development both need software engineering approach, project management skills, and tools and techniques but
with a different approach. Programming projects are easy to implement. Why? Because there we need to focus
only on what is to be done. How it is to be done, is taken care by inbuilt code, which is pre tested for quality. In
data warehousing projects, it becomes important to determine how things are to be done due to unavailability of
standardized approach. We have tried to standardize the procedure of data warehousing to improve success rate
of data warehousing projects.
Nature of queries of operational system and data warehousing are different from each other.
Operational system queries report business data which is current or of recent past, and is generally atomic in
nature. On the other hand, data warehousing queries report result of business process analysis over a long period
of time span either in a single time slot or in a sequence of multiple time spans. In the absence of data
warehousing solution, performance assessment of various services provided, and their effects are collected by
means of carrying out large scale public surveys to get public feedback. In case of data warehousing solution,
getting such information becomes automatic with the help of running suitable database queries over developed
data warehousing solution. These are major benefits of implementing data warehousing solutions.
Organization of this paper is as follows. Section 2 covers literature review. Section 3 discusses design
of data warehouse development standardization framework (DWDSF). Section 4 gives details of National Rural
Health Mission (NRHM) project and its data warehousing solution. Section 5 includes some of the business
manager’s query expectations, SQL queries for their requirements and their results. Section 6 concludes the
paper.
II. Literature Review
Data warehouse failure can be attributed to issues of data quality, database schema, quality of modeling
tool, and quality of operational system [4]. Misunderstood user expectations, lacking in understanding the data,
adopting an over complicated architecture for the solution, considering user defined requirements as final and
complete, and always blaming quality of data for data warehouse failure are key reasons of data warehouse
project’s failure [6]. [3] Highlighted the problem of lack of standardization of data warehouse development
methodologies. Organizations often struggle with successfully delivering a data warehouse solution due to
execution challenges, and poor planning which could lead to years of extended wait times before organization
can realize the benefit of the data warehouse [6]. Gartner Says More Than 50 Percent of Data Warehouse
Projects Will Have Limited Acceptance or Will Be Failures Through 2007 as a result of a lack of attention to
data quality issues, according to Gartner, Inc [2]. "Consistency and accuracy of data is critical to success with
BI, and data quality must be viewed as a business issue and responsibility, not just an IT problem" [2]. "It is
hard to believe that IT organizations still build data warehouses with little or no business involvement," said
Frank Buytendijk, research vice president at Gartner"[2]. Failure rate of data warehousing projects has been a
major concern in data warehousing industry. Data warehousing is an essential element of decision support. It
aims at enabling the knowledge user to make better and faster daily business decisions [1]. The process of data
warehousing starts by analyzing operational system of the organization. Incomplete and poor operational system
design is bound to put hurdles in smooth design and implementation of data warehousing solution [7].
3. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 31 | Page
Research contributions in the field of data warehousing could be classified in the following two broad
categories.
1. Developing tools and technologies to support data warehouse implementation.
2. Developing methodologies and procedural steps to be used by data warehouse developers to help them in
effective implementation of data warehousing solutions.
III. Design of Data Warehouse Development Standardization Framework (DWDSF)
For accomplishing any challenging task following factors are important: i. Availability of advance
tools and technologies. Ii. Knowledge of appropriate approach to be followed. If we consider implementation of
data warehousing solutions, we find that advance tools and technologies are available. We only need to identify
and follow the approach that would be effective for successful implementation of data warehousing solution.
We have designed a framework named “Data Warehouse Development Standardization Framework” (DWDSF),
to help data warehouse developer’s community in implementing effective data warehousing solutions. Our
framework (DWDSF) proves highly useful in helping out in deciding on such issues of appropriate approach
selection. DWDSF framework specifies i. Steps of data warehouse development process, ii.Decisions to be
taken during each step, and iii.Approach to be followed, during each step.
Data warehouse development steps proposed in our framework take care of managerial aspects of data
warehouse implementation, whereas details of activities proposed during each of the steps in the framework take
care of technical issues of data warehouse implementation.
Our framework’s methodology is based on two layer approach. Layer 1 describes macro steps of the
data warehouse development process. Layer-2 describes micro steps under each macro step and its determinants.
Layer-1 develops general understanding of data warehouse development approach to be followed whereas layer
2 ensures that each of the steps is undertaken with a clear idea of its purpose. Two-layer approach ensures that
we strictly adhere to the proposed framework and at each of the micro step check if we are getting the
mentioned outcome of that micro step. Macro layer guides about broad steps to be followed for data warehouse
development where as micro layer of the framework guides about specific atomic level activities to be
performed during each of the macro step and describes its purpose in terms of outcome. Our framework is “what
and how” kind of framework which not only tells the steps to be followed but also describes the way to do each
step. The major focus of our framework is on requirement analysis and data warehouse design phase, because
these two are highly complex tasks in data warehousing projects. Implementation of designed system is
relatively simple, involving routine activities of physical implementation. Implementation involves execution of
ETL code or steps to clean and load data into data warehouse schema.
Our framework suggests following basic steps (Macro Steps) to be followed for data warehouse
implementation:
1. Understand goals and objectives
2. Identify measurements to be made that could assess if goals and objectives are met
3. Investigate data items required for above measurements
4. Design schema to be populated with the identified data items
5. Iterate above steps till complete system is designed.
6. Implement designed system and populate with data.
Many research papers have repeatedly given steps to be followed for data warehouse solution
development but internal details of steps are very brief. We have tried to dig deep into Micro details of steps that
are useful for data warehouse developer’s community. We have given details of the micro layer of the
framework in Table 3 which are represented in the form of a Purpose table for data warehouse development
process explaining purpose of each atomic step.
Table 3: Purpose table for data warehouse development process explaining purpose of each atomic step
Sr. No. Data warehouse development atomic step Purpose of the atomic step
1. Identify Goals and Objectives Determines reporting needs
2. Identify Key Business Processes Determines facts to be measured and dimensions of
analysis
3. Identify Organizational Structure Determine different business users
4. Investigate functionality supported in the
operational system
Determines quality of the operational system
Two-way quality assessment is performed i. At its
own, ii. Quality assessment for data warehouse
implementation
5. Analyze reports generated by the operational
system
Determines quality of the operational system
6. Identify reports the operational system could
not generate
Determines limitations of the operational system
7. Interact with different users groups Determines quality of operational system, users
expectations from data warehousing solution
4. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 32 | Page
8. Analyze operational system application forms
and database schema
Determines data items available in the operational
system and data items that are missing
9. Analyze query types business managers want
to run
Determines data warehouse schema design details,
and grain of measurements to be taken in data
warehouse schema design
10. Analyze parameters of business process
analysis
Determines dimension tables and their attributes
11. Analyze key time intervals to be taken in
most of the reporting
Determines execution schedule of ETL for
population and updating of data warehouse schema
12. Design dimensional model and design ETL
code
Determines dimension tables, fact tables, their
attributes, and schedules and inputs of ETL code
execution
13. Create tables, run ETL code, Completes data warehouse schema implementation
14. Run queries to generate reports to test data
warehouse
Tests data warehouse for its performance and
determines its quality
Data warehouse development challenges are due to multiple data sources, multiple data technologies,
high volume, lack of user awareness about the data warehouse technology and its uses, design decisions for facts
and dimension tables, about de-normalization or snow-flaking or fact table constellation, decision of data grain
selection, architecture selection for implementing data warehouse, selection of suitable materialized views,
design of ETL code, and complex data warehouse requirement analysis.
“Problem is not with the technology. It is with the approach we follow”. Therefore our framework is based on
procedural aspect of data warehousing. It is related to the subject of software engineering therefore we have
taken standard SDLC and have tried to customize it for data warehousing projects.
Techniques applied for system investigation:
In our framework we have proposed following methods of system investigation:
1. Questionnaire distribution to business managers to collect their response.
2. Personal interview of business managers.
3. Organizational document study.
4. Analysis of operational system input forms and output reports.
5. Users query expectations.
Basic steps of DWDSF are given in table no. 4.
Table 4: Basic steps of DWDSF
Sr. No. Data warehouse Implementation Steps
1. Preliminary Investigation
2. Assessment of Operational System
3. Feasibility test report
4. Analysis of needs
5. Data warehouse design
6. Data warehouse implementation
3.1 Features of DWDSF Framework
1. Integrated approach has been developed which guides in every aspect of data warehousing practices or data
warehouse project execution.
2. The proposed Framework is based on procedural aspects of data warehousing that helps data warehouse
developers to follow specific approach and ensure expected outcome from any project
3. Knowledge of physical implementation and the physical implementation methods used by a specific data
warehousing tool helps in deciding which tool to use in a particular case but cannot govern developers
about what steps to follow and what approach to follow during each step. This is where our framework
proves highly useful.
3.2 Fitness test of operational system and its significance in data warehouse implementation
Signs of failure are “deciding to proceed for data warehousing with unfit operational system”.
Whereas, very first sign of success is “deciding to proceed for data warehousing only if operational system is fit
for data warehousing”. Taking decisions of data warehouse implementation without provisioning for current
limitations leads to an attempt that most likely fails in the end. If operational system is not fit for data
warehousing, we can suggest improvements in the operational system and postpone the data warehouse
implementation. If only to fulfil commercial interests of development organisation, the project is undertaken, it
is likely to fail. We have worked to investigate fitness of operational system for data warehouse
implementation, and offer details of improvements required in the operational system if it is not fit for the data
warehousing.
5. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 33 | Page
3.3 Causes of Heterogeneity
1. Non uniform data formats or structures
2. Change of technology with time
Non uniform data formats or structures are manageable with precautions in operational systems
whereas changes in technology with time are unavoidable and need to be managed during data warehouse
design. No prior research suggests in the direction of improving operational system design improvements. Most
of the works are based on basic assumption that operational systems in use are fit for data warehouse
implementation. Our approach of work is different from others. Our Framework suggests that the first step of
data warehouse implementation is assessment of operational system.
3.4 Feasibility analysis for Data warehousing
1. Is operational system strong enough to support data warehousing? yes or no if no, then feasibility test
result is big no, for data warehousing.
2. Need of analytical assessment of business? yes or no if answer is no then feasibility test is big no.
Managing multiple technologies of data sources is a requirement for data warehouse implementation. One
method to deal with this requirement is to use data format standardization standard called XML. Diagrammatic
representation of the model for heterogeneity management is as follows:
Fig No.-1: dealing with data heterogeneity using XML data standard.
IV. National Rural Health Mission and its Data warehousing solution
NRHM started with HMIS reporting system and later introduced MCTS application for mother and child care
programme but without considering National level workload and infrastructure requirements. IT implementation
goals and vision was not clearly defined.
NRHM goals for mother and child care module are as follows:
1. Reduce mother mortality rate (MMR)
2. Reduce infant mortality rate (IMR)
3. Reduce total fertility rate (TFR)
4. Prevention and reduction of anaemia in women aged between 15 to 49 years
5. Increase institutional delivery
MMR, IMR, TFR, prevention of anaemia, institutional delivery all are related to pregnant women
health care during and post pregnancy. From the partial goal list, we are able to identify that we need to
maintain data related to pregnant women registration and routine checkups details during and post pregnancy.
By referring the health domain terminology from available organisational documents, we have identified
following business processes
1. pregnant women ante natal care (ANC)
2. pregnant women delivery
3. pregnant women post natal care (PNC)
From these business processes, tentative facts are mother registration, mother checkups, mother delivery, mother
PNC checkups details.
In addition to HMIS, and MCTS, NRHM project implemented State level websites which are maintained by
each state separately.
We thoroughly investigated all operational systems implemented for NRHM and identified some useful
points of concern. In depth investigation revealed that HMIS is a reporting system from facility centres to
controlling body in a specific file format. It is actually a mix of manual data collection, compilation and
computerized reporting platform and is not a full-fledged operational system. It involves both manual and
computerized intervention to produce aggregated state level and national level reports. HMIS was developed to
monitor various health programs for which fund is being released to all states, districts and so on but it is not
able to support data warehouse designing, primarily because it does not maintain atomic data. It does not contain
all data items which are required to analyze its business operations for strategic decision making. HMIS does
not support any business process implementation and also is not able to generate any report on user request. It
6. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 34 | Page
does not have any user interface to take requests and generate customized reports. It is just a data collection
platform from Central monitoring and evaluation division at national level, which then compiles the collected
data and publishes compiled reports that are for user reference at any level.There are many data items which are
not important for operational system but are very useful and are required for performance analysis. Such data
items need to be included in operational system to facilitate effective data warehousing solution development.
For example age of pregnant women or patient, address, socio economic status and other demographical data
elements are not useful for operational system but are highly useful for demographic evaluation of operation’s
performance. Operational system needs or requirements and analytical system’s needs or requirements are quite
different. But analytical systems are developed from operational systems therefore needs for analytical system
also need to be investigated for operational system development, to facilitate analytical application development
of future. It is generally not considered, resulting in difficulties faced during development of analytical solutions
like data warehouses. Data is not available for cases that received services from outside agencies (absence of
external data).
Although, the operational system design is one of the most critical factor deciding success or failure of
a data warehousing solution but most of the data warehousing research directly considers designing data
warehouse without focusing much on operational system design and implementation. We have considered a step
of quality assessment of operational system and its role in data warehouse designing. We have also discovered
its solution which is based on a unique idea of designing operational system and Data warehousing Solution in
an integrated manner to take care of objectives of both types of systems. We have followed an approach to first
identify how much satisfied users are with the existing operational system and what problems they have been
facing with their current operational system and why should they go for Data warehousing Solution and whether
they need a data warehousing solution or need to improve their operational system first.Organizations size is
big, volume of transaction is big, the age of project is also sufficient, need of complex analysis is also required.
All above findings pass feasibility test of data warehousing implementation. But if the operational system is not
systematically implemented it needs to be analyzed and improved so that, future Data warehousing Solution will
also be a successful implementation. Why to design a Data warehousing Solution that will fail in future? First
step should be to identify modifications required in the operational system for data warehouse design and re-
design and implement it accordingly instead of going for a Data warehousing Solution. Can we say a project like
NRHM does not require Data warehousing Solution? if no, is NRHM project’s computerization in the current
form suitable to support data warehouse implementation? if no., then what is next solution or option? Answer is
to investigate operational system of NRHM, then do the required implementation changes in the operational
system and then either sequentially or in parallel, proceed for data warehousing implementation. If we you look
into Indian governments scenario, most of the projects despite being at National level are not fully automated. In
such situations, real reporting is not possible. It in turn would result in decision taking which is not based on real
facts. Why government schemes prove inefficient? Because, they are not able to pin point the problems in the
absence of effective operational system.
Table 5: Information useful in size estimation
Sr. No. Element Description Element’s Count at National Level
1. Community Health Centres (CHCs) 5396
2. Primary Health Centres (PHCs) 25308
3. Health Sub Centres (SCs) 153655
4. States 36
5. Doctors 23089
6. Ayush Doctors 11925
7. Ayush Paramedicos 4785
7. Auxillary Nurse Midwife (ANM) 70891
8. Accredited social health activists (ASHA) 890000
Table 5 contains information useful in size estimation of data warehousing solution for NRHM project.
HMIS focused on reporting figures without considering parameters responsible for these figures. MCTS also has
some of the required parameters missing. We have implemented ETL simulation code for generating various
measurement counts based on inputs received from business managers and populating data warehouse schema.
It was done due to non availability of atomic data in the operational system. Sample simulation code is given in
figure number 2.
import java.sql.*;
public class ANCRegistrationPopulate {
static final String JDBC_DRIVER = "oracle.jdbc.driver.OracleDriver";
static final String DB_URL = "jdbc:oracle:thin:@localhost:1521";
static final String USER = "d1";
static final String PASS = "d1";
7. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 35 | Page
public static void main(String[] args) {
int darc_Id=1;
int loopStart=1;
int loopEnd=3941;
Connection conn = null;
PreparedStatement prSt = null;
int k = 0; //m_h_p
int l = 0; //m_p_p
int m = 0; //m_s_e_s
int n = 0; //anm
int p = 0; //asha
int q = 0; //registration_count
int recordCount=0;
try {
Class.forName("oracle.jdbc.driver.OracleDriver");
System.out.println("Connecting to a selected database...");
conn = DriverManager.getConnection(DB_URL, USER, PASS);
System.out.println("Connected database successfully...");
System.out.println("Inserting records into the table...");
String sql = "INSERT INTO ANC_Registration VALUES(?,?,?,?,?,?,?,?,?)";
prSt = conn.prepareStatement(sql);
for (int i = loopStart; i <= loopEnd; i++) {
for (int j = 1; j <= 9687; j++) {
prSt.setInt(1, darc_Id);
darc_Id++; //check shift appropriately
prSt.setInt(2, i);
prSt.setInt(3, j);
while (!(k >= 1 && k <= 3620)) {
k = (int) (Math.random() * 36200);
}
prSt.setInt(4, k);
k = 0;
while (!(l >= 1 && l <= 1222)) {
l = (int) (Math.random() * 12220);
}
prSt.setInt(5, l);
l = 0;
while (!(m >= 1 && m <= 292608)) {
m = (int) (Math.random() * 2926080);
}
prSt.setInt(6, m);
m = 0;
while (!(n >= 1 && n <= 113696)) {
n = (int) (Math.random() * 1136960);
}
prSt.setInt(7, n);
n = 0;
while (!(p >= 1 && p <= 113696)) {
p = (int) (Math.random() * 1136960);
}
prSt.setInt(8, p);
p = 0;
if ((j > 0 && j <= 1489) || (j > 1495 && j <= 1516) || (j > 2892 && j <= 2895) || (j > 2897 && j <=
8797)) {
q = (int) (Math.random() * 40);
} else if ((j > 1489 && j <= 1495) || (j > 2534 && j <= 2544)) {
q = (int) (Math.random() * 120);
} else if ((j > 2639 && j <= 2875) || (j > 9235 && j <= 9668) || (j > 1578 && j <= 2533)) {
q = (int) (Math.random() * 300);
8. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 36 | Page
} else if ((j > 1516 && j <= 1573) || (j == 2534) || (j > 2544 && j <= 2546) || (j > 2895 && j <=
2897) || (j == 9235) || (j == 9687) || (j > 8797 && j <= 8811)) {
q = (int) (Math.random() * 900);
} else if ((j > 1573 && j <= 1578) || (j > 2546 && j <= 2575) || (j > 2575 && j <= 2639) || (j > 2875
&& j <= 2872) || (j > 9668 && j <= 9686)) {
q = (int) (Math.random() * 2500);
} else if ((j > 8811 && j <= 9234)) {
q = 0;
} else {
q = 1;
}
prSt.setInt(9, q);
q = 0;
prSt.executeUpdate();
recordCount++;
}
}
System.out.println("Inserted "+recordCount+"records into the table...");
} catch (SQLException se) {
se.printStackTrace();
} catch (Exception e) {
e.printStackTrace();
} finally {
try {
if (prSt != null) {
conn.close();
}
} catch (SQLException se) {
}// do nothing
try {
if (conn != null) {
conn.close();
}
} catch (SQLException se) {
se.printStackTrace();
}//end finally try
}//end try
System.out.println("Goodbye!");
}//end main
}
Figure no. - 2, simulation code for generating various measurements counts
V. Results
For testing implemented solution, we tried to run queries requested by the business managers. We executed
many queries requested by business managers and could generate the required reports as per user expectations.
Some sample requested reports, executed SQL query, and the corresponding output is given as follows:
Query Objective:
Display number of women registered for ANC based on mother’s education and family income.
SQL Statement:
Selectmother_education,family_income,sum(total_number_of_women_register)
"number_of_women_registered" from anc_registration b, mother_socioeconomic_status a where
b.MOTHER_SOCIO_ECONOMIC_STATUS_I=a.MOTHER_SOCIOECONOMIC_STATUS_ID group by
mother_education,family_income;
Output Result:
9. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 37 | Page
MOTHER_EDUCATION FAMILY_INCOME number_of_women_registe
red
Illiterate Less Than Five Hundred 26496005
Illiterate Five Hundred to Two Thousand 26252838
Illiterate Two Thousand to Five Thousand 26231282
Illiterate Five Thousand to Ten Thousand 26193811
Illiterate More than Ten Thousand 26046384
Literate without School Less Than Five Hundred 65100652
Literate without School Five Hundred to Two Thousand 64831334
Literate without School Two Thousand to Five Thousand 64777949
Literate without School Five Thousand to Ten Thousand 64647611
Literate without School More than Ten Thousand 65008987
Primary Less Than Five Hundred 65110166
Primary Five Hundred to Two Thousand 64611857
Primary Two Thousand to Five Thousand 64610229
Primary Five Thousand to Ten Thousand 64900781
Primary More than Ten Thousand 64749036
Junior High School Less Than Five Hundred 38386477
Junior High School Five Hundred to Two Thousand 38623918
Junior High School Two Thousand to Five Thousand 38391916
Junior High School Five Thousand to Ten Thousand 38568355
Junior High School More than Ten Thousand 38783664
Query Objective:
Display number of women registered for ANC based on mother’s age range.
SQL Statement:
select age_range,sum(total_number_of_women_register)
“TOTAL_NUMBER_OF_WOMEN_REGISTERED”from anc_registration b, mother_health_profile a where
b.mother_health_profile_id=a.mother_health_profile_id group by age_range;
Output Result:
AGE_RANGE TOTAL_NUMBER_OF_WOMEN_REGISTERED
Fifteen to Ninteen 358787552
Twenty to Twenty Four 356999924
Twenty Five to Twenty Nine 275261056
Thirty to Thirty Four 248242433
Thirty Five to Thirty Nine 126079298
Forty to Forty Four 59300773
Query Objective:
Display number of women registered for ANC based on distance from home to facility centre.
SQL Statement:
select distance_from_home_to_facility,sum(total_number_of_women_register)”
TOTAL_NUMBER_OF_WOMEN_REGISTER” from anc_registration b, mother_health_profile a where
b.mother_health_profile_id=a.mother_health_profile_id group by distance_from_home_to_facility; Output
Result:
DISTANCE_FROM_HOME_TO_FACILITY TOTAL_NUMBER_OF_WOMEN_REGISTERED
With in Three Kilometres 284720925
Between Three to Six Kilometres 284912481
Between Six to Ten Kilometres 285097249
Between Ten to Twenty Kilometres 285036463
More than Twenty Kilometres 284903918
Query Objective:
Display number of women registered for ANC based on year wise registration on each facility centre.
SQL Statement:
select a.facility_centre_name,c.year_of_Day,sum(total_number_of_women_register) from anc_registration b,
facility_centre_dimension a, Date_Dimension c where b.facility_centreid=a.facility_centre_id and
c.date_id=b.date_id group by facility_centre_name,c.year_of_Day;
Output Result:
10. Data Warehouse Development Standardization Framework (DWDSF): A Way to Handle Data
DOI: 10.9790/0661-1901022938 www.iosrjournals.org 38 | Page
FACILITY_CENTRE_NAME YEAR_OF_DAY TOTAL_NUMBER_OF_WOMEN_REGISTERED
Karaiyur 2005 40028
Athimanjeripettai 2005 39944
Negamam 2005 42282
Nallattipalayam 2005 42959
Madukkarai 2005 41086
Vadalur 2005 42920
Thoppur 2005 40247
Theerthamalai 2005 40320
Acharapakkam 2005 40047
Velliyanai 2005 40983
Manmangalam 2005 41311
NPHC Goberaha 2006 7435
Salt 2006 7023
Bahadrabad 2006 7255
Yameshwor 2006 6963
Rikhnikhal 2006 7278
Dugdda 2006 7328
Badalu 2006 7304
Chamba 2006 7179
Dwida 2006 7251
Almora Sub-divisional Hospital 2006 638119
Irumpedu 2007 7325
Desur 2007 7002
Chithathoor 2007 7033
Veppalodai 2007 6734
We have designed and implemented data warehouse on the basis of analysis of only one program of
NRHM project out of 17 programs it is running. It is like a prototype which verifies quality of the developed the
solution, gets feedback of user group to move further with design and development of data warehouse for other
programs as well. The results generated by our solution were highly satisfactory. We could meet out all of user’s
reporting expectations which is highly encouraging result.
VI. Conclusion
Review of literature on data warehousing revealed issues of high failure rates of data warehousing
projects. In this paper, we analyzed possible reasons of data warehousing project failure and found that most of
the reasons are related to procedural aspects. We designed a framework called Data Warehouse Development
Standardization Framework (DWDSF) which is based on procedural aspects of data warehousing, has tried to
standardize steps of data warehousing and elaborates ways to undertake each of the steps. We have also
discussed role of operational system in data warehousing and its feasibility test to identify its suitability for
proceeding with data warehousing project or to proceed for improvement of the implemented operational system
first. We also proposed an integrated approach of solution development where requirements of both operational
system and data warehousing solutions could be considered together to achieve quality in both type of solutions
and avoiding effect of poor operational system in data warehousing solution. We implemented data warehousing
solution for one of the key subject areas of NRHM project and executed some useful queries over the developed
solution and found that all the queries executed satisfactorily and produced results as per expectations of the
business managers. Our implementation worked as a solution prototype and produced an early feedback of the
users to proceed further with more implementation.
Declaration: The sole purpose of our work is analysis of the technical and managerial aspects of data warehousing. Material
referred, names, terminology, and contents referred from concerned external materials are properties of respective owners. We have used
them purely for the academic enhancement and not for any other purpose or intentions. Our objective is not to criticize any system. We have
tried to investigate and suggest possible improvements.
References
[1]. Rami Rifaieh, Nabila Aïcha Benharkat, “Query-based Data Warehousing Tool”, DOLAP’02, November 8, 2002, McLean, Virginia,
USA. Copyright 2002 ACM 1-58113-590-4/02/0011
[2]. STAMFORD, CONN February 24, 2005
[3]. Jukic, N., (2006), “Modeling Strategies And Alternatives For Data Warehousing Projects.”, Communications of The ACM April
[4]. 2006/Vol. 49, No. 4, p-83-90
[5]. http://www.embarcadero.com,Tech Notes, “Why Data Warehouse Projects Fail”, January 2008.
[6]. Predicts2012,”business_intelligence_still_subject_to_non-technical_challenges”
http://www.saiindia.gov.in/english/home/our_products/Audit_Report/Government_Wise/union_audit/recent_reports/union_perform
ance/2009_2010/Civil_%20Performance_Audits/Report_no_8/chap_9.pdf
[7]. Chris JenningsInsights, “Five Key Causes on Why Data Warehouse Projects Fail”, December 22, 2014
[8]. Deepak Asrani, Renu Jain, & Usha Saxena, “Limitations of Operational System Design and Requirements of a Successful Data
Warehousing Solution: A Gap Analysis for NRHM”, 4th International Conference on Emerging Trends in Engineering and
Technology College Of Engineering ,Teerthanker Mahaveer University, 24-25 Apr, 2015