Dynamic Hyper-Converged Future Proof Your Data CenterDataCore Software
IT organizations are continuously striving to reduce the amount of time and effort to deploy new resources for the business. Data center and remote office infrastructures are often complex and rigid to deploy, causing operational delays. As a result, many IT organizations are looking at a hyper-converged infrastructure.
Read this whitepaper to discover that a hyper-converged approach is flexible and easy to deploy and offers:
• Lower CAPEX because of lower up-front prices for infrastructure
• Lower OPEX through reductions in operational expenses and personnel
• Faster time-to-value for new business needs
Achieve New Heights with Modern AnalyticsSense Corp
Businesses can leverage modern cloud platforms and practices for net-new solutions and to enhance existing capabilities, resulting in an upgrade in quality, increased speed-to-market, global deployment capability at scale, and improved cost transparency.
In this webinar, Josh Rachner, data practice lead at Sense Corp, will help prepare you for your analytics transformation and explore how to make the most on new platforms by:
Building a strong understanding of the rise, value, and direction of cloud analytics
Exploring the difference between modern and legacy systems, the Big Three technologies, and different implementation scenarios
Sharing the nine things you need to know as you reach for the clouds
You’ll leave with our pre-flight checklist to ensure your organization will achieve new heights.
Agile Testing Days 2017 Introducing AgileBI SustainablyRaphael Branger
"We now do Agile BI too” is often heard in todays BI community. But can you really "create" agile in Business Intelligence projects? This presentation shows that Agile BI doesn't necessarily start with the introduction of an iterative project approach. An organisation is well advised to establish first the necessary foundations in regards to organisation, business and technology in order to become capable of an iterative, incremental project approach in the BI domain. In this session you learn which building blocks you need to consider. In addition you will see what a meaningful sequence to these building blocks is. Selected aspects like test automation, BI specific design patterns as well as the Disciplined Agile Framework will be explained in more and practical details.
The Pivotal Business Data Lake provides a flexible blueprint to meet your business's future information and analytics needs while avoiding the pitfalls of typical EDW implementations. Pivotal’s products will help you overcome challenges like reconciling corporate and local needs, providing real-time access to all types of data, integrating data from multiple sources and in multiple formats, and supporting ad hoc analysis.
Pervasive analytics through data & analytic centricityCloudera, Inc.
Cloudera and Teradata discuss the best-in-class solution enabling companies to put data and analytics at the center of their strategy, achieve the highest forms of agility, while reducing the costs and complexity of their current environment.
Dynamic Hyper-Converged Future Proof Your Data CenterDataCore Software
IT organizations are continuously striving to reduce the amount of time and effort to deploy new resources for the business. Data center and remote office infrastructures are often complex and rigid to deploy, causing operational delays. As a result, many IT organizations are looking at a hyper-converged infrastructure.
Read this whitepaper to discover that a hyper-converged approach is flexible and easy to deploy and offers:
• Lower CAPEX because of lower up-front prices for infrastructure
• Lower OPEX through reductions in operational expenses and personnel
• Faster time-to-value for new business needs
Achieve New Heights with Modern AnalyticsSense Corp
Businesses can leverage modern cloud platforms and practices for net-new solutions and to enhance existing capabilities, resulting in an upgrade in quality, increased speed-to-market, global deployment capability at scale, and improved cost transparency.
In this webinar, Josh Rachner, data practice lead at Sense Corp, will help prepare you for your analytics transformation and explore how to make the most on new platforms by:
Building a strong understanding of the rise, value, and direction of cloud analytics
Exploring the difference between modern and legacy systems, the Big Three technologies, and different implementation scenarios
Sharing the nine things you need to know as you reach for the clouds
You’ll leave with our pre-flight checklist to ensure your organization will achieve new heights.
Agile Testing Days 2017 Introducing AgileBI SustainablyRaphael Branger
"We now do Agile BI too” is often heard in todays BI community. But can you really "create" agile in Business Intelligence projects? This presentation shows that Agile BI doesn't necessarily start with the introduction of an iterative project approach. An organisation is well advised to establish first the necessary foundations in regards to organisation, business and technology in order to become capable of an iterative, incremental project approach in the BI domain. In this session you learn which building blocks you need to consider. In addition you will see what a meaningful sequence to these building blocks is. Selected aspects like test automation, BI specific design patterns as well as the Disciplined Agile Framework will be explained in more and practical details.
The Pivotal Business Data Lake provides a flexible blueprint to meet your business's future information and analytics needs while avoiding the pitfalls of typical EDW implementations. Pivotal’s products will help you overcome challenges like reconciling corporate and local needs, providing real-time access to all types of data, integrating data from multiple sources and in multiple formats, and supporting ad hoc analysis.
Pervasive analytics through data & analytic centricityCloudera, Inc.
Cloudera and Teradata discuss the best-in-class solution enabling companies to put data and analytics at the center of their strategy, achieve the highest forms of agility, while reducing the costs and complexity of their current environment.
Enterprise Data Management - Data Lake - A PerspectiveSaurav Mukherjee
This document discusses the evolution of the enterprise data management over the years, the challenges of the current CTOs and chief enterprise architects, and the concept of the Data Lake as a means to tackle such challenges. It also talks about some reference architectures and recommended tool set in today’s context.
Making the Case for Hadoop in a Large Enterprise-British AirwaysDataWorks Summit
Making the Case for Hadoop in a Large Enterprise
British Airways
Alan Spanos
Data Exploitation Manager
British Airways
Jay Aubby
Architect
British Airways
The CSC Big Data Analytics Insights service enables clients who do not have an analytics capability to implement the business, data and technology changes to gain business benefit from an initial set of analytics based on a roadmap of changes created by CSC or provided from a compatible set of inputs.
CSC Analytic Insights Implementation has four phases:
Stage 1: Analytic Engagement
Stage 2: Analytic Discovery
Stage 3: Implementation Planning
Stage 4: Embedding Analysis .
Matt Aslett (The451Group) and Deirdre Mahon (RainStor) examine the evolving data management landscape and how RainStor's Online Data Retention (OLDR) repository fits into the equation.
** Watch the video to accompany these slides: https://www.cloverdx.com/webinars/starting-your-modern-dataops-journey **
- What is "Data Ops" and why should you consider it?
- How to begin your transition to a DevOps and DataOps-style of work
- How agile methodologies, version control, continuous integration or 'infrastructure as code' can improve the effectivity of your teams
- How you can use technology like CloverDX to start with DataOps
Discover how to make your development and data analytics processes more efficient and effective by shifting to a Dev/DataOps approach.
More CloverDX webinars: https://www.cloverdx.com/webinars
Twitter: https://twitter.com/cloverdx
LinkedIn: https://www.linkedin.com/company/cloverdx/
Get a free 45 day trial of the CloverDX Data Management Platform: https://www.cloverdx.com/trial-platform
The Business Data Lake is a new approach to information management, analytics and reporting that better matches the culture of business and better enables organizations to truly leverage the value of their information.
THE FUTURE OF DATA: PROVISIONING ANALYTICS-READY DATA AT SPEEDwebwinkelvakdag
Data lakes & data warehouses, whether on-premises or in the cloud promise to provide a centralized, cost-effective and scalable foundation for modern analytics. However, organisations continue to struggle to deliver accurate, current and analytics-ready data sets in a timely fashion. Traditional ingestion tools weren’t designed to handle hundreds or even thousands of data sources and the lack of lineage forces data consumers to manually aggregate information from sources they trust. In this session, you’ll learn how to future-proof your modern data environment to meet the needs of the business for the long term. We'll examine how to overcome common challenges, the related must-have technology solutions in the data lake/ data warehousing world, using real-world success stories and even a few architecture tips from industry experts.
Extended Data Warehouse - A New Data Architecture for Modern BI with Claudia ...Denodo
This presentation has been extracted from a full webinar organized by Denodo. To learn more click here: http://bit.ly/1FOMD90
Big Data, Internet of Things, Data Lakes, Streaming Analytics, Machine Learning… these are just a few of the buzzwords being thrown around in the world of data management today. They provide us with new sources of data, new forms of analytics, and new ways of storing, managing and utilizing our data. The reality however, is that traditional Data Warehouse architectures are no longer able to handle many of these new technologies and a new data architecture is required.
So what does the new architecture look like? Does the enterprise data warehouse still have a role? Where do these new technologies fit in? How can business users easily and quickly access the various sources of data and analytic results at the right time to make the right decisions in this new world order?
Dr. Claudia Imhoff addresses these questions and presents the Extended Data Warehouse architecture (XDW), demonstrating the need for each component and how an enterprise combines these into appropriate workflows for proper decision support.
Hadoop 2.0 - Solving the Data Quality ChallengeInside Analysis
The Briefing Room with Dr. Claudia Imhoff and RedPoint Global
Live Webcast on July 22, 2014
Watch the archive:
https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=7bb4cbc33402c3b5f649343052cb9a6d
Whether data is big or small, quality remains the critical characteristic. While traditional approaches to cleansing data have made strides, nonetheless, data quality remains a serious hurdle for all organizations. This is especially true for identity resolution in customer data, but also for a range of other data sets, including social, supply chain, financial and other domains. One of the most promising approaches for solving this decades-old challenge incorporates the power of massive parallel processing, a la Hadoop.
Register for this episode of The Briefing Room to learn from veteran Analyst Dr. Claudia Imhoff, who will explain how Hadoop 2.0 and its YARN architecture can make a serious impact on the previously intractable problem of data quality. She’ll be briefed by George Corugedo of RedPoint Global, who will show how his company’s platform can serve as a super-charged marshaling area for accessing, cleansing and delivering high-quality data. He’ll explain how RedPoint was one of the first applications to be certified for running on YARN, which is the latest rendition of the now-ubiquitous Hadoop.
Visit InsideAnlaysis.com for more information.
Data warehouses have become a popular mechanism for collecting, organizing, and making information readily available for strategic decision making. The ability to review historical trends and monitor near real-time operational data has become a key competitive advantage for many organizations. Yet the methods for assuring the quality of these valuable assets are quite different from those of transactional systems. Ensuring that the appropriate testing is performed is a major challenge for many enterprises. Geoff Horne has led a number of data warehouse testing projects in both the telecommunications and ERP sectors. Join Geoff as he shares his approaches and experiences, focusing on the key “uniques” of data warehouse testing including methods for assuring data completeness, monitoring data transformations, and measuring quality. He also explores the opportunities for test automation as part of the data warehouse process, describing how it can be harnessed to streamline and minimize overhead.
Data summit connect fall 2020 - rise of data opsRyan Gross
Data governance teams attempt to apply manual control at various points for consistency and quality of the data. By thinking of our machine learning data pipelines as compilers that convert data into executable functions and leveraging data version control, data governance and engineering teams can engineer the data together, filing bugs against data versions, applying quality control checks to the data compilers, and other activities. This talk illustrates how innovations are poised to drive process and cultural changes to data governance, leading to order-of-magnitude improvements.
Creating a Successful DataOps Framework for Your Business.pdfEnov8
As data is universally important and has a major role in decision-making and other business operations, a strong data-driven culture has become extremely important for business organizations.
This calls for a successful and efficient DataOps framework. Let us explore more about this emerging methodology.
Enterprise Data Management - Data Lake - A PerspectiveSaurav Mukherjee
This document discusses the evolution of the enterprise data management over the years, the challenges of the current CTOs and chief enterprise architects, and the concept of the Data Lake as a means to tackle such challenges. It also talks about some reference architectures and recommended tool set in today’s context.
Making the Case for Hadoop in a Large Enterprise-British AirwaysDataWorks Summit
Making the Case for Hadoop in a Large Enterprise
British Airways
Alan Spanos
Data Exploitation Manager
British Airways
Jay Aubby
Architect
British Airways
The CSC Big Data Analytics Insights service enables clients who do not have an analytics capability to implement the business, data and technology changes to gain business benefit from an initial set of analytics based on a roadmap of changes created by CSC or provided from a compatible set of inputs.
CSC Analytic Insights Implementation has four phases:
Stage 1: Analytic Engagement
Stage 2: Analytic Discovery
Stage 3: Implementation Planning
Stage 4: Embedding Analysis .
Matt Aslett (The451Group) and Deirdre Mahon (RainStor) examine the evolving data management landscape and how RainStor's Online Data Retention (OLDR) repository fits into the equation.
** Watch the video to accompany these slides: https://www.cloverdx.com/webinars/starting-your-modern-dataops-journey **
- What is "Data Ops" and why should you consider it?
- How to begin your transition to a DevOps and DataOps-style of work
- How agile methodologies, version control, continuous integration or 'infrastructure as code' can improve the effectivity of your teams
- How you can use technology like CloverDX to start with DataOps
Discover how to make your development and data analytics processes more efficient and effective by shifting to a Dev/DataOps approach.
More CloverDX webinars: https://www.cloverdx.com/webinars
Twitter: https://twitter.com/cloverdx
LinkedIn: https://www.linkedin.com/company/cloverdx/
Get a free 45 day trial of the CloverDX Data Management Platform: https://www.cloverdx.com/trial-platform
The Business Data Lake is a new approach to information management, analytics and reporting that better matches the culture of business and better enables organizations to truly leverage the value of their information.
THE FUTURE OF DATA: PROVISIONING ANALYTICS-READY DATA AT SPEEDwebwinkelvakdag
Data lakes & data warehouses, whether on-premises or in the cloud promise to provide a centralized, cost-effective and scalable foundation for modern analytics. However, organisations continue to struggle to deliver accurate, current and analytics-ready data sets in a timely fashion. Traditional ingestion tools weren’t designed to handle hundreds or even thousands of data sources and the lack of lineage forces data consumers to manually aggregate information from sources they trust. In this session, you’ll learn how to future-proof your modern data environment to meet the needs of the business for the long term. We'll examine how to overcome common challenges, the related must-have technology solutions in the data lake/ data warehousing world, using real-world success stories and even a few architecture tips from industry experts.
Extended Data Warehouse - A New Data Architecture for Modern BI with Claudia ...Denodo
This presentation has been extracted from a full webinar organized by Denodo. To learn more click here: http://bit.ly/1FOMD90
Big Data, Internet of Things, Data Lakes, Streaming Analytics, Machine Learning… these are just a few of the buzzwords being thrown around in the world of data management today. They provide us with new sources of data, new forms of analytics, and new ways of storing, managing and utilizing our data. The reality however, is that traditional Data Warehouse architectures are no longer able to handle many of these new technologies and a new data architecture is required.
So what does the new architecture look like? Does the enterprise data warehouse still have a role? Where do these new technologies fit in? How can business users easily and quickly access the various sources of data and analytic results at the right time to make the right decisions in this new world order?
Dr. Claudia Imhoff addresses these questions and presents the Extended Data Warehouse architecture (XDW), demonstrating the need for each component and how an enterprise combines these into appropriate workflows for proper decision support.
Hadoop 2.0 - Solving the Data Quality ChallengeInside Analysis
The Briefing Room with Dr. Claudia Imhoff and RedPoint Global
Live Webcast on July 22, 2014
Watch the archive:
https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=7bb4cbc33402c3b5f649343052cb9a6d
Whether data is big or small, quality remains the critical characteristic. While traditional approaches to cleansing data have made strides, nonetheless, data quality remains a serious hurdle for all organizations. This is especially true for identity resolution in customer data, but also for a range of other data sets, including social, supply chain, financial and other domains. One of the most promising approaches for solving this decades-old challenge incorporates the power of massive parallel processing, a la Hadoop.
Register for this episode of The Briefing Room to learn from veteran Analyst Dr. Claudia Imhoff, who will explain how Hadoop 2.0 and its YARN architecture can make a serious impact on the previously intractable problem of data quality. She’ll be briefed by George Corugedo of RedPoint Global, who will show how his company’s platform can serve as a super-charged marshaling area for accessing, cleansing and delivering high-quality data. He’ll explain how RedPoint was one of the first applications to be certified for running on YARN, which is the latest rendition of the now-ubiquitous Hadoop.
Visit InsideAnlaysis.com for more information.
Data warehouses have become a popular mechanism for collecting, organizing, and making information readily available for strategic decision making. The ability to review historical trends and monitor near real-time operational data has become a key competitive advantage for many organizations. Yet the methods for assuring the quality of these valuable assets are quite different from those of transactional systems. Ensuring that the appropriate testing is performed is a major challenge for many enterprises. Geoff Horne has led a number of data warehouse testing projects in both the telecommunications and ERP sectors. Join Geoff as he shares his approaches and experiences, focusing on the key “uniques” of data warehouse testing including methods for assuring data completeness, monitoring data transformations, and measuring quality. He also explores the opportunities for test automation as part of the data warehouse process, describing how it can be harnessed to streamline and minimize overhead.
Data summit connect fall 2020 - rise of data opsRyan Gross
Data governance teams attempt to apply manual control at various points for consistency and quality of the data. By thinking of our machine learning data pipelines as compilers that convert data into executable functions and leveraging data version control, data governance and engineering teams can engineer the data together, filing bugs against data versions, applying quality control checks to the data compilers, and other activities. This talk illustrates how innovations are poised to drive process and cultural changes to data governance, leading to order-of-magnitude improvements.
Creating a Successful DataOps Framework for Your Business.pdfEnov8
As data is universally important and has a major role in decision-making and other business operations, a strong data-driven culture has become extremely important for business organizations.
This calls for a successful and efficient DataOps framework. Let us explore more about this emerging methodology.
CSC - Presentation at Hortonworks Booth - Strata 2014Hortonworks
Come hear about how companies are kick-starting their big data projects without having to find good people, hire them, and get IT to prioritize it to get your project off the ground. Remove risk from your project, ensure scalability , and pay for just the nodes you use in a monthly utility pricing model. Worried about Data Governance, Security, want it in the cloud, can’t have it in the cloud….eliminate the hurdles with a fully managed service backed by CSC. Get your modern data architecture up and running in as little as 30 days with the Big Data Platform As A Service offering from CSC. Computer Science Corporation is a Certified Technology Partner of Hortonworks and is a Global System Integrator with over 80,000 employees globally.
Leveraging Cloud for Non-Production EnvironmentsCognizant
Moving to the cloud not only enables application development and testing organizations to reduce capital outlays; it can also reduce IT cycle times while improving quality.
In simple words, DataOps is all about aligning the way you manage your data with the objectives you have for that data. Let’s know in detail what actually DataOps is!
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida CLARA CAMPROVIN
Análisis empresariales cuando los necesite, en cualquier lugar
Jet Enterprise es una solución de inteligencia empresarial y generación de informes desarrollada específicamente para satisfacer las necesidades propias de los usuarios de Microsoft Dynamics. Ahora puede juntar toda su información en un mismo lugar y permitir que quien usted quiera de la organización realice fácilmente sofisticados análisis empresariales desde cualquier sitio. Capacite a los usuarios para tomar mejores decisiones, más rápido, prácticamente con cualquier dispositivo.
Con Jet Enterprise dispone de:
Una solución completa de inteligencia empresarial y generación de informes, lista para usar en solo 2 horas
Más de 80 paneles y plantillas de informes
7 cubos pregenerados personalizables
Un almacén de datos
Integración directa con sus datos de Microsoft Dynamics y posibilidad de conectarse a otros sistemas empresariales pertinentes
Posibilidad de crear paneles en cuestión de minutos, sin necesidad de conocer la estructura de datos subyacente
Jet Mobile opcional, para acceder a sus datos desde cualquier sitio a través de un navegador web o un dispositivo móvil
Una plataforma robusta de automatización y personalización del almacenamiento de datos
«Comenzamos con datos de Sage Pro, datos de NAV 2009 y, además, datos incorporados de la nueva empresa que habíamos adquirido, por lo que ahora estamos usando tres sistemas de datos. Las ventajas de combinar los tres sistemas en Jet Enterprise han sido enormes».
– Davis & Shirtliff
Éxito inmediato = rápido ROI y bajo coste de propiedad
Muchas soluciones de inteligencia empresarial conllevan costes ocultos, como implementaciones prolongadas y difíciles, personalizaciones caras y precio elevado de las licencias cuando se amplían a un gran número de usuarios. Jet Enterprise se suele instalar en unas dos horas, requiere un nivel mínimo de formación de los usuarios y ofrece licencias para un número ilimitado de usuarios. Los usuarios habitualmente experimentan un incremento de los ingresos brutos en los primeros 12 meses de uso.
Data Lake-based Approaches to Regulatory-Driven Technology ChallengesBooz Allen Hamilton
Booz Allen Hamilton has found that a data lake-based approach to CA3 requirements is scalable, extensible, and improves the range and sophistication of analyses that can be supported while providing higher levels of data control and security.
Appliance Warehouse Service Plan.The discussion focuses on the.docxfestockton
Appliance Warehouse Service Plan.
The discussion focuses on the appliance Warehouse Service Plan that is made up of the testing plan, an implementation plan and the training plan for the sake of the bettering of services in a warehouse. The testing plan is meant to manage the systems through QA standards meeting the needs of the customers. The implementation plan elaborates and indicates whether one should use parallel, direct, phased, or pilot changeover strategies. The training plan, on the other hand, indicates what a training plan would include for affected employees, such as appointment setters, technicians, management, and the parts department.
Testing Plan
The main reason for the testing plan is to validate and verify the information from the main source or the end to end target warehouse. The two major testing plans for include program testing and acceptance testing (Lewis, 2017). The plan should verify the following, the business required documents, ETL design for the documents, sources to target on the mapping process and the data model for the source and the target schemas. The documents that are considered are meant for the ETL development process in the testing plan. The testing plan is meant further for the supervisors or the quality analysis team to confirm that the work is concerning the objective of the organization. The process of testing might also include the configuration management system and the data quality validation and verification process.
Implementation Plan
The plan for the implementation of the systems is the same as the process that is considered during the development process of the entire system to meet the goals of the organization. The steps to consider for the whole plan of the implementation include the analysis and the enhancement requests, the writing of very simplified and new programs, restructuring of the database, analysis of the program library and its cost, and the reengineering of the test program. The first phase parallels the analysis phase as the parallel strategy is considered for the entire process, which entails the analysis phase of the SDLC. The steps two to four process entails the combining and the construction activities that are done on a new system majorly on a small scale. The last step is meant to parallel the testing that is commonly done during the implementation process. The testing process ensures that the process is free of risk as a quality assurance process (Liang & Hui, 2016).
Training Plan
The training plan should be made up of a training matrix in which it will guide them to know who needs the training what they need from the training and why they want the training not forgetting when they need the training(Kwak,2016). The matrix will allow for the planning and the preparation for the training avoiding scrambling when the due date for the training comes around. The requirements are automatically updated when the employees get done with the first training before transferri ...
Appliance Warehouse Service Plan.The discussion focuses on the.docxRAHUL126667
Appliance Warehouse Service Plan.
The discussion focuses on the appliance Warehouse Service Plan that is made up of the testing plan, an implementation plan and the training plan for the sake of the bettering of services in a warehouse. The testing plan is meant to manage the systems through QA standards meeting the needs of the customers. The implementation plan elaborates and indicates whether one should use parallel, direct, phased, or pilot changeover strategies. The training plan, on the other hand, indicates what a training plan would include for affected employees, such as appointment setters, technicians, management, and the parts department.
Testing Plan
The main reason for the testing plan is to validate and verify the information from the main source or the end to end target warehouse. The two major testing plans for include program testing and acceptance testing (Lewis, 2017). The plan should verify the following, the business required documents, ETL design for the documents, sources to target on the mapping process and the data model for the source and the target schemas. The documents that are considered are meant for the ETL development process in the testing plan. The testing plan is meant further for the supervisors or the quality analysis team to confirm that the work is concerning the objective of the organization. The process of testing might also include the configuration management system and the data quality validation and verification process.
Implementation Plan
The plan for the implementation of the systems is the same as the process that is considered during the development process of the entire system to meet the goals of the organization. The steps to consider for the whole plan of the implementation include the analysis and the enhancement requests, the writing of very simplified and new programs, restructuring of the database, analysis of the program library and its cost, and the reengineering of the test program. The first phase parallels the analysis phase as the parallel strategy is considered for the entire process, which entails the analysis phase of the SDLC. The steps two to four process entails the combining and the construction activities that are done on a new system majorly on a small scale. The last step is meant to parallel the testing that is commonly done during the implementation process. The testing process ensures that the process is free of risk as a quality assurance process (Liang & Hui, 2016).
Training Plan
The training plan should be made up of a training matrix in which it will guide them to know who needs the training what they need from the training and why they want the training not forgetting when they need the training(Kwak,2016). The matrix will allow for the planning and the preparation for the training avoiding scrambling when the due date for the training comes around. The requirements are automatically updated when the employees get done with the first training before transferri.
This presentation was provided by Daniel Calto of Elsevier during the NISO virtual conference, Research Information Systems: The Connections Enabling Collaboration, held on August 16, 2017.
Accelerate Innovation & Productivity With Rapid Prototyping & Development - ...Attivio
Today, development teams typically need hundreds of person hours to develop an application or to fully
integrate a new platform. Prototypes and Proofs of Concept (PoC) also take many weeks (or even months)
to develop. If you could significantly reduce these timeframes, you would accelerate time to market and
expedite PoCs and rollouts. This advantage saves money and reduces the risk of missing features, late deliveries or inadequate testing.
December 2015 - TDWI Checklist Report - Seven Best Practices for Adapting DWA
1. c1 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
TDWI CHECKLIST REPORT
TDWI RESEARCH
tdwi.org
Seven Best Practices for
Adopting Data Warehouse
Automation
By David Loshin
Sponsored by:
3. 2 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
The word agile conveys a number of meanings. Its simple definition
as an adjective is “quick and well-coordinated,” but the term has
taken on additional meaning in the context of system design and
development, largely in terms of rapid development, increased
partnership among information technology (IT) and business partners,
and leveraging teamwork to achieve short-term objectives that build
toward solving more complex business challenges.
Applying agile development methodologies to data warehousing
promises a number of benefits: quicker development and movement
into production, faster time to value, reduced start-up and overhead
costs, and simplified access for business users. Yet there are some
prerequisites to transitioning to using the agile approach for data
warehousing:
• Evaluate the existing environment to identify the best
opportunities for achieving the benefits of adoption
• Establish good practices for leveraging more agile technologies
where possible
• Identify the right technologies to facilitate that transition
There are a number of technologies that accelerate design and
development, improve cycle time in producing reports and analyses,
and enhance the IT-business collaboration. Some are platform
oriented, such as columnar databases, in-memory computing, and
Hadoop, which all seek to leverage faster performance to improve
analytical results. Alternatively, data warehouse automation
(DWA) tools blend user requirements and repeatable processes
to automatically generate the components of a data warehouse
environment. These tools simplify the end-to-end production of a
data warehouse, encompassing the entire development life cycle,
including source system analysis, design, development, generation
of data integration scripts, building, deployment, generation of
documentation, testing, support for ongoing operations, impact
analysis, and change management.
We are rapidly moving away from the monolithic, single-system
enterprise data warehouse and toward a hybrid environment that
uses the most appropriate technologies to address specific data
challenges. That environment will encompass many components and
will benefit from reduced complexity through the use of tools like DWA.
This checklist discusses seven practices for determining the value
proposition of adopting DWA and establishing the foundation that will
ease its adoption.
FOREWORD
The success of the conventional IT management approach to data
warehousing and business intelligence (BI) has opened the door for
a growing population of data warehouse consumers, both within and
outside of the organization. Although many of these consumers’ needs
are met by existing reports or by providing access for straightforward
queries, a combination of factors has created a bottleneck in rapidly
addressing business-user demands. Many business analysts are
becoming more sophisticated in their investigations, requiring
additional consulting from their IT counterparts to develop data
extracts and reports.
At the same time, though, IT budgets and staffing remain constrained.
The result is that scheduling limited IT consulting resources elongates
the time from when a business user requests a data product to the
time when that data product is delivered (“cycle time”).
There are two main risks of long cycle times. First, the data product
is delivered after the window of opportunity for taking advantage
of its results. Second, frustrated users may abandon the use of the
enterprise data warehouse and adopt their own “shadow” reporting
and analytics tools and methods, bypassing any governance
procedures intended to ensure enterprisewide consistency.
Long cycle times and development bottlenecks lead to missed
business opportunities. Therefore, increasing agility by eliminating
those bottlenecks will increase the benefits of your reporting and
analytics investment.
To find where those bottlenecks hinder productivity, analyze the end-
to-end process for satisfying BI and analysis requests. It is likely that
the source of the logjam is in the design-develop-test loop for new
reports and extracts; of course, automation tools can reduce or even
eliminate the development blockage and speed time to value.
Analyzing the report development cycle time provides a benchmark for
the time and resources required (in general) to satisfy business needs.
This benchmark can be used as the starting point for optimizing
existing processes as well as providing a metric for evaluation of how
DWA tools can speed time to value.
ANALYZE THE CYCLE TIME FOR SATISFYING
BUSINESS REQUESTS
NUMBER ONE
4. 3 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
The excitement over new technologies can sometimes overwhelm
common sense. Organizations have invested significant amounts
of money and staff effort in architecting, developing, and
productionalizing their existing data warehouse and BI platforms. It
would be surprising, risky, and generally unwise for any organization
to completely rip out a trusted legacy data warehousing platform
and replace it with systems generated using DWA tools or move
directly to cloud-hosted data warehousing providers.
A more conservative approach would embrace a transition
strategy that incrementally migrates data warehousing and BI
applications from legacy platforms to more agile environments.
Start off by establishing an innovation lab within the existing
environment in which new techniques can be piloted and evaluated
for interoperability with existing systems. This allows different
approaches to be considered while constraining the alternatives to
ones that are compatible with the production environment.
Design a hybrid data warehousing environment architecture
that accommodates the introduction of new technologies and
emerging agile development paradigms while maintaining the
production operations of established conventional systems. Employ
virtualization methods to provide layers on top of existing platforms,
and abstract and differentiate the functionality from
its implementation.
Designing a hybrid architecture provides the flexibility to integrate
different data warehousing approaches that have already been
vetted. In turn, adopting the right technology mix allows the
application teams to develop facades abstracting the underlying
capabilities. This allows for reengineering without disrupting
production systems, enables dual operations during a testing period
to verify that the new approaches are trustworthy, and facilitates
seamless transitions for the user community when the time comes to
migrate applications.
Adjustments to the data warehousing tools and platform
infrastructure can potentially reduce development bottlenecks and
speed time to value. This decision triggers the evaluation of vendors
with products that purport to deliver on that promise, selecting
candidate alternatives, and choosing one (or more) to integrate
within the enterprise.
Balance the benefits and costs of using existing data warehouse
systems versus introducing newer technology. Recognize that
introducing new technology into the organization requires more than
just purchasing the license and installing the tool. It also requires
design and development time, an integration effort, training to
empower product users, and a communications plan to transition
legacy users to modernized platforms. Any plan for data warehouse
environment modernization must incorporate the cost and resources
needed to support those tasks to assess the potential return on
investment and to compare and contrast candidate technologies.
Develop a value model for comparing data warehousing alternatives
both against the existing platform as well as against each other. The
key is to select the right variables for comparison that will lead to
greater agility, lower costs, and better outcomes. Some variables for
comparison include:
• Application development complexity. How easy is it to design,
configure, and deploy a new data warehouse or add new subject
areas within an existing data warehouse environment?
• Application development time. How long does it take for a data
warehouse to become operational, focusing on the end-to-end
process of design, development, and implementation?
• Skills requirements. What types of skills are required by the
development team members and how long does it take to acquire
those skills?
• Report development turnaround time. How long is the cycle
time for new reports?
• End-user ease of use. How easy is it to empower data
consumers to use the developed data warehouse without IT
support?
• Resource requirements. What are the necessary resources for
implementation?
• Cost. What are the accumulated start-up and operational costs?
This value model will provide quantitative comparisons to guide
technology selection and is likely to highlight the value of DWA tools.
DEVISE A VALUE MODEL FOR COMPARING DATA
WAREHOUSING ALTERNATIVES
ARCHITECT A HYBRID ENVIRONMENT
NUMBER TWO NUMBER THREE
5. 4 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
Our world has evolved into one where new, diverse sources of
massive amounts of data are emerging every day. From an analytics
perspective, the impact of the explosion of data sets originating
outside the typical enterprise is twofold. On the one hand, there is a
growing availability of information that can inform internal subject
area profiles, such as enhancing customer behavior information
by analyzing streaming social media posts. On the other hand, the
broad variety of both structured and unstructured data formats
creates significant complexity for the data warehouse professionals
who are unaccustomed to programming with Web services and APIs.
The conventional approach to data warehousing focuses on ingesting
one or two data sets at a time, originating from sources inside
the organization. However, as the speed, volume, and diversity of
externally sourced data grow, the organization must move faster
in absorbing data sources and putting those data streams to
productive use.
Integrating new data sources requires that:
• The incoming data set is profiled as part of a discovery process
• Representative models are designed for data ingestion
• The incoming data model elements are mapped to existing data
warehouse data elements
• Definitions are captured (or inferred) to ensure semantic
consistency between new data sources and existing data models
• The data warehouse model must be augmented to absorb any
newly defined data elements of interest from the new data source
• Transformations are programmed to align incoming data with the
corresponding data elements in the existing warehouse model
• The incoming data sets are continuously monitored to verify that
the data exchange interface has not changed
Taking these steps prior to ingestion requires effective project
management, but ensuring the scheduling, operations, and fidelity
of production intake processes may overwhelm those choosing
to manually oversee the tasks. Data warehouse automation
simplifies these processes by automating discovery and alignment
of data source metadata with existing warehouse metadata and
orchestrating data ingestion, transformation, and loading. The
generated utilities are effectively self-documenting, enabling the
developers to understand what the generated code does.
Adopting the agile development methodology for data warehouse
development suggests that the days of the IT data practitioner
as the data warehouse gatekeeper are over. Business users are
significantly more knowledgeable than they were during the early
days of data warehousing, and they have a much lower dependence
on IT staff to meet many of their more mundane needs when it
comes to developing reports or performing simple queries.
Yet, as more enlightened business analysts devise sophisticated
analyses, the burden on IT staff only increases beyond the typical
development, operations, and maintenance of the data warehouse
platforms. Adopting DWA tools to support building and managing
data warehouses reduces the IT staff’s burden for design and
development. This provides more time for IT staff members to focus
on helping business users evaluate their specific business problems
and on how reporting and analysis can solve those problems (rather
than delivering reports in a virtual vacuum).
According to the Agile Alliance, the agile software development
methodology emphasizes “close collaboration between the
programmer team and business experts; face-to-face communication
(as more efficient than written documentation); frequent delivery
of new deployable business value; tight, self-organizing teams.”1
Fostering increased developer-user communication can reduce the
cycle time for developing reports and more complex analyses. Such
communication can also lead to increased user satisfaction.
Train your data professionals to actively engage business users,
effectively solicit business requirements, and (together with the
business users) translate those requirements into directives to
formulate reports and analyses that actually meet business needs.
By relying on tools that can automate data discovery, warehouse
modeling, and report creation, the collaborative knowledge transfer
between IT staff and their business partners will educate the
business analysts to become more self-sufficient. Business-user
self-sufficiency triggers a virtuous cycle by yet again reducing the
IT burden, freeing those resources to work with other business
users to engender more self-sufficiency to provide ever more
sophisticated analytics.
BE AGILE IN INGESTING NEW DATA SOURCES TRAIN DATA ENGINEERS TO COLLABORATE WITH
BUSINESS USERS
NUMBER FOUR NUMBER FIVE
1
“What Is Agile Software Development?” downloaded November 6, 2015 from
http://www.agilealliance.org/the-alliance/what-is-agile.
6. 5 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
The notion of a business intelligence competency center (BICC) is
motivated by the early challenges associated with data warehouse
development and deployment, specifically around data integration,
population of the data warehouse, and coordination among the
business users to ensure data quality, consistency, and leveraging
the investment in data warehouse environment technologies. The
original goal of the BICC was to focus on standardizing policies for
data warehouse use, centralizing the support of BI, and developing
good practices and repeatable processes across the organization.
Practically, though, in many environments two conflicting facets
hampered the success of the BICC. The first is the difficulty in
organizing an IT team intended to enforce policies across business
function boundaries. The second is the constraints on flexibility
imposed as a way to enforce data usage policies and standards.
As a result, the BICC often becomes the bottleneck as decisions
about technology and architecture are subsumed within demands for
report development and configuration. Essentially, strong centralized
oversight reduces agility in the organization, limiting business-user
gains from unfettered data exploration and discovery.
As organizations look to more agile methods to more quickly
allow business users to control their data use, it may be time to
reevaluate the goals of the BICC, determine where artificial barriers
to knowledge have been institutionalized, and assess ways that it
can be adapted to meet the needs of an increasingly enlightened
business community.
Examine how to revise the BICC’s charter to differentiate between
operational decision making associated with day-to-day tactical
management of the data warehouse environment and strategic
decision making related to evolving the environment over time. Adopt
guiding principles that expand data warehouse utilization, such
as simplifying the organizational data warehousing architecture,
acquiring tools that reduce overall time to value, simplifying
the repeatable processes for data warehouse instantiation, and
increasing hands-on training and knowledge transfer with business
partners. As business use increases, encourage business experts
to take on a role within the structure of the BICC so that its
management and oversight can be diffused among all stakeholders.
If becoming more agile means encouraging closer collaboration
between developers and business experts, we can take that a
step further in the data warehousing and BI world to enable the
transfer of skills from the IT professionals to the business experts to
make them self-sufficient. In effect, creating a more collaborative
environment between the IT/data staff and the business users
increases business-user independence and facilitates increased
information utilization.
At the same time, empowering self-sufficient business users
reduces business dependence on IT staff, reduces IT costs to
support enhanced data use, and frees IT resources to focus on BI
functionality, improved analytical precision, and expanded analytical
services.
Self-service BI leverages intuitive user interfaces and data
accessibility functionality that both guide the user in designing
reports and analyses and exercise controls over what can and
cannot be accessed. In addition, self-service BI relies on the
availability of a semantic metadata repository that lists business
terms and table column names and provides a shared glossary to
ensure consistency in use when formulating new reports.
One of the key benefits of self-service BI is that, as the business
users become more adept at developing their own analyses,
they can speed the review cycle for discovery of actionable
knowledge. However, supporting this decreased cycle time requires
complementary speed of data warehouse development. Agile tools
such as DWA can be used to quickly implement the warehouses and
marts that provide flexibility to end users in configuring their own
reports while limiting access to data needing additional protection.
RETHINK THE ROLE OF THE BI COMPETENCY CENTERSUPPORT BUSINESS SELF-SERVICE
NUMBER SEVEN NUMBER SIX
7. 6 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
Business analysts have acquired a more cultivated awareness of
data discovery, data preparation, report formulation, and predictive
analytics and are increasingly taking control of their own reports
and analyses. These same data consumers are benefitting from the
explosion of data sources that can be captured within the analytics
environment. However, as the number of data sources increases
and the data sets become more diverse, maintaining a competitive
advantage hinges on the speed at which new data sources can be
acquired, ingested, and prepared for discovery and analysis.
It is unlikely that any organization will completely rip out their
existing data warehouses and replace them with any specific new
technology. Instead, the future data warehouse environment will
be an evolving hybrid environment composed of conventional data
warehouse architectures and high-performance components layered
on the Hadoop ecosystem, and it will perhaps include specialty
appliances or software accelerators such as columnar databases
and in-memory computing systems.
As this evolution proceeds, data warehouse automation tools help to
maintain agility while supporting the demands of legacy customers.
The items on this checklist help in preparing your organization for
defining assessment criteria, evaluating the existing environment,
determining where there are opportunities for agility, and adopting
DWA tools as a way to accelerate the transformation of the
enterprise analytics environment.
When evaluating alternatives, the tools guide the developers in
source analysis, in capturing requirements, and in automatically
generating the data integration, loading, and presentation
components of a data warehouse. The next step is to develop an
integration plan. Pilot the tools by focusing on developing a data
warehouse to support a specific business function or to analyze
a specific subject area (like customers or vendors). Align your
development methodology with the way that developers utilize the
tools and the ways users expect to use the resulting warehouse.
Reflect on lessons learned in terms of rapid design cycles,
empowering users with self-service capabilities, and ingesting new
data sources. These lessons will inform repeatable processes that
will guide the evolution of the future data warehouse environment.
CONSIDERATIONS
DEVELOPING A FEASIBLE INTEGRATION PLAN FOR
DATA WAREHOUSE AUTOMATION
8. 7 TDWI RESEARCH tdwi.org
TDWI CHECKLIST REPORT: SEVEN BEST PRACTICES FOR ADOPTING DATA WAREHOUSE AUTOMATION
TDWI Research provides research and advice for data professionals
worldwide. TDWI Research focuses exclusively on business
intelligence, data warehousing, and analytics issues and teams
up with industry thought leaders and practitioners to deliver both
broad and deep understanding of the business and technical
challenges surrounding the deployment and use of business
intelligence, data warehousing, and analytics solutions. TDWI
Research offers in-depth research reports, commentary, inquiry
services, and topical conferences as well as strategic planning
services to user and vendor organizations.
ABOUT TDWI RESEARCH
ABOUT THE AUTHOR
David Loshin, president of Knowledge Integrity, Inc., (www.
knowledge-integrity.com), is a recognized thought leader, TDWI
instructor, and expert consultant in the areas of data management
and business intelligence. David is a prolific author regarding
business intelligence best practices, as the author of numerous
books and papers on data management, including Big Data
Analytics: From Strategic Planning to Enterprise Integration with
Tools, Techniques, NoSQL, and Graph and The Practitioner’s Guide
to Data Quality Improvement, with additional content provided at
www.dataqualitybook.com. David is a frequent invited speaker at
conferences, Web seminars, and sponsored websites and channels
including TechTarget and The Bloor Group. His best-selling book
Master Data Management has been endorsed by many data
management industry leaders.
David can be reached at loshin@knowledge-integrity.com.
TDWI Checklist Reports provide an overview of success factors for
a specific project in business intelligence, data warehousing, or
a related data management discipline. Companies may use this
overview to get organized before beginning a project or to identify
goals and areas of improvement for current projects.
ABOUT TDWI CHECKLIST REPORTS
timextender.com
TimeXtender is a world-leading data warehouse automation vendor
dedicated to Microsoft SQL Server.
The TimeXtender software, TX DWA, revolutionizes the way a data
warehouse is developed and maintained by automating all manual
data warehouse processes—from design to development, operation,
and maintenance to change management. TimeXtender ensures an
improved and inexpensive solution that is fully documented.
The TX DWA software enables medium and large data-driven
enterprises to get business intelligence done faster, more efficiently,
and with less stress by providing “one truth” to improve decision-
making processes and overall business performance that reduces
costs and saves valuable time. TX DWA turns a business intelligence
project into a business intelligence process with a flexible solution
that expands as the business evolves and grows.
TimeXtender collaborates with VAR and OEM partners across six
continents, providing more than 2,600 customers in 61 countries
all over the world with its advanced data warehouse automation
software.
Why wait days before taking action? With TimeXtender, data is
available at your fingertips in mere hours!
ABOUT OUR SPONSOR