This document provides information about quality management system procedures including forms, tools, and strategies. It discusses developing procedures according to ISO 9001 standards and simplifying procedures. Quality management tools like check sheets, control charts, Pareto charts, scatter plots, Ishikawa diagrams, and histograms are explained. Additional related topics like quality management systems, courses, and standards are also listed.
The document outlines Iron Mountain Legal Discovery's efforts to improve service delivery quality from 2009-2011, including establishing quality control processes, implementing a Six Sigma measurement framework, consulting on process improvement, and achieving higher customer satisfaction scores and productivity through training and audits. Key achievements included reducing rework costs and improving first-time resolution rates and overall quality above 95%. Future plans include expanding the service delivery quality scope and aligning more closely with engineering, products, and global standards.
This document discusses using SharePoint for quality management systems (QMS) in regulated environments like medical device manufacturing. It provides an example of a company currently using shared network drives and document control software that wants to implement SharePoint to help with collaboration, processes, and document management while complying with FDA and ISO quality standards. The document asks if anyone has case studies or expertise in customizing SharePoint for these specific QMS needs. It also provides an overview of common quality management tools like check sheets, control charts, Pareto charts, scatter plots, Ishikawa diagrams, histograms and their uses in quality control.
The document describes Riskpro, an organization that provides risk management consulting services to companies in India. It offers services such as risk assessment, process improvement, training and workshops, and knowledge management programs to clients in the ITES/BPO/KPO industries. Some key risks it identifies for these industries include high attrition, data security and privacy issues, fraud risk, regulatory non-compliance, and concentration risk. The document provides examples of offerings like web-based training and solutions to address various risks faced by ITES/BPO companies.
A presentation about the added value of combining qualitative and quantitative methods. It begins with a brief discussion of qualitative research and how it is distinct from yet shares basic principles with quantitative research, followed by a discussion of four important ways mixed methods -- integrating qualitative and quantitative -- adds value to our research efforts, and then a discussion of mixed methods research -- what it is, typologies, alternatives to typologies, and the use of diagrams.
The document provides an overview of the Avoca Quality Consortium, which was formed to develop best practices for proactive quality management of outsourced clinical trials. It discusses the state of the clinical trials industry that led to the Consortium's formation, including increased outsourcing and globalization. It then outlines the Consortium's approach, deliverables produced to date like Quality Agreements and metrics, and areas of ongoing focus such as guidelines for effective quality oversight and operationalizing proactive quality management. The overall goal is to improve quality and efficiency in clinical trials through collaboration and standardization.
Process asset library as process improvement and knowledge sharing toolKobi Vider
This document discusses process asset libraries (PALs), including definitions, approaches, and references in the Capability Maturity Model Integration (CMMI). It defines a PAL as a library used to store and share process assets across an organization. Process assets can include documents, templates, lessons learned, and other artifacts. The document explains that CMMI references PALs and emphasizes their importance in establishing standard processes, collecting improvement information, and enabling consistent process performance and organizational learning.
2004 E2M - The ShopView Story Information Package.PDFMelissa Jones
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help boost feelings of calmness and well-being.
This document provides information about quality management system procedures including forms, tools, and strategies. It discusses developing procedures according to ISO 9001 standards and simplifying procedures. Quality management tools like check sheets, control charts, Pareto charts, scatter plots, Ishikawa diagrams, and histograms are explained. Additional related topics like quality management systems, courses, and standards are also listed.
The document outlines Iron Mountain Legal Discovery's efforts to improve service delivery quality from 2009-2011, including establishing quality control processes, implementing a Six Sigma measurement framework, consulting on process improvement, and achieving higher customer satisfaction scores and productivity through training and audits. Key achievements included reducing rework costs and improving first-time resolution rates and overall quality above 95%. Future plans include expanding the service delivery quality scope and aligning more closely with engineering, products, and global standards.
This document discusses using SharePoint for quality management systems (QMS) in regulated environments like medical device manufacturing. It provides an example of a company currently using shared network drives and document control software that wants to implement SharePoint to help with collaboration, processes, and document management while complying with FDA and ISO quality standards. The document asks if anyone has case studies or expertise in customizing SharePoint for these specific QMS needs. It also provides an overview of common quality management tools like check sheets, control charts, Pareto charts, scatter plots, Ishikawa diagrams, histograms and their uses in quality control.
The document describes Riskpro, an organization that provides risk management consulting services to companies in India. It offers services such as risk assessment, process improvement, training and workshops, and knowledge management programs to clients in the ITES/BPO/KPO industries. Some key risks it identifies for these industries include high attrition, data security and privacy issues, fraud risk, regulatory non-compliance, and concentration risk. The document provides examples of offerings like web-based training and solutions to address various risks faced by ITES/BPO companies.
A presentation about the added value of combining qualitative and quantitative methods. It begins with a brief discussion of qualitative research and how it is distinct from yet shares basic principles with quantitative research, followed by a discussion of four important ways mixed methods -- integrating qualitative and quantitative -- adds value to our research efforts, and then a discussion of mixed methods research -- what it is, typologies, alternatives to typologies, and the use of diagrams.
The document provides an overview of the Avoca Quality Consortium, which was formed to develop best practices for proactive quality management of outsourced clinical trials. It discusses the state of the clinical trials industry that led to the Consortium's formation, including increased outsourcing and globalization. It then outlines the Consortium's approach, deliverables produced to date like Quality Agreements and metrics, and areas of ongoing focus such as guidelines for effective quality oversight and operationalizing proactive quality management. The overall goal is to improve quality and efficiency in clinical trials through collaboration and standardization.
Process asset library as process improvement and knowledge sharing toolKobi Vider
This document discusses process asset libraries (PALs), including definitions, approaches, and references in the Capability Maturity Model Integration (CMMI). It defines a PAL as a library used to store and share process assets across an organization. Process assets can include documents, templates, lessons learned, and other artifacts. The document explains that CMMI references PALs and emphasizes their importance in establishing standard processes, collecting improvement information, and enabling consistent process performance and organizational learning.
2004 E2M - The ShopView Story Information Package.PDFMelissa Jones
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help boost feelings of calmness and well-being.
QMS SharePoint Wireframe - download and edit for you useMelissa Jones
This document outlines the proposed site architecture and navigation for a new SharePoint QMS site. It includes wireframes for the site pages, proposed navigation structure with top-level and sub-menus, and assignments for the project team to create lists, libraries, and pages according to the wireframes and then test and deploy the new site.
Part 3 - SharePoint QMS Anyone Can Make - Data DictionaryMelissa Jones
This document provides guidance on setting up lists and libraries in a specific order for a SharePoint quality management system (QMS). It begins by listing standard reference lists that provide quality standards information. It then lists foundational lists for organizing work centers and job descriptions. Core QMS lists and libraries are listed next, including the main document library which can link to other lists. Optional lists for areas like training and customer feedback are also included. Each list and library is then described in more detail regarding its purpose and design.
Jim McCall produced a quality framework model for the US Air Force to bridge the gap between users and developers. The framework defines quality factors divided into software quality, product operation, product revision, and product transition categories. McCall's triangle of quality relates these factors to quality metrics. Product operation factors are defined by metrics expressions to quantify attributes. The approach is user-oriented at the highest level and software-oriented at lower levels, allowing periodic quantification during development.
Metadata Quality Assurance Framework at QQML2016 conference - full versionPéter Király
This document presents a Metadata Quality Assurance Framework to measure and improve metadata quality. It analyzes typical metadata issues like non-informative fields and proposes measuring structural elements like completeness, cardinality, uniqueness, and language specification to predict record quality. Metrics are defined using a problem catalog of known issues mapped to discovery scenarios. Visualizations of early measurement results are shown to identify outliers and inform metadata improvements. The framework is intended to be scalable, transparent, and collaborative.
Quality measurement - How to measure the quality of any object?Grzegorz Grela
THE FRAMEWORK OF QUALITY MEASUREMENT
Quality is the degree to which a set of inherent characteristics fulfils requirements. (ISO 9000)
Requirements and inherent characteristics create finite sets.
Requirements may have both different importance and different values depending on who formulates them.
Requirements do not have to be constant in time.
Quality measurement may be conducted on two levels: analytical and synthetic.
Source: Grela, G. (2015). The Framework of Quality Measurement. Management (18544223), 10(2).
http://www.fm-kp.si/zalozba/ISSN/1854-4231/10_177-191.pdf
15 Months to Certification: Using SharePoint as the Platform for an ISO 9001 ...Barry Peters
Telerx implemented SharePoint to achieve ISO 9001 certification within 15 months. Key aspects included using SharePoint for document control, change control, internal audits, and tracking non-conformances and corrective actions. Workflows automated document approval and status updates. Record control leveraged SharePoint information management policies. The system provided audit trails and version control while supporting continuous improvement processes required by ISO 9001.
The Planning Quality Framework is a collection of tools and techniques that use planning data to help councils understand their development management service performance and benchmark against others. It involves quantitative data like application counts and approval rates, as well as qualitative customer surveys. The framework provides regular reports to give councils insights into the value and quality of their work. It is a low-effort way to focus improvement efforts compared to traditional benchmarking approaches.
The document describes a Quality Management System calibration records list on SharePoint. It provides instructions for accessing calibration records, viewing item details and attachments, running monthly calibration due reports, and exporting the records list to Excel. Maintaining up-to-date calibration records is important for regulatory compliance, and corrective actions may be issued for overdue equipment.
This document summarizes quality assurance and quality control processes. It discusses that quality control focuses on identifying defects in finished products through reactive testing, while quality assurance aims to prevent defects through a proactive process focus. It then outlines the implementation of quality assurance in various areas including stores, production, packing, and batch storage and release. Key activities involve inspection planning, issuing control numbers, monitoring processes, sampling, testing, and record keeping. The goals are to control waste, enhance product quality consistency, and improve customer faith.
The presentation in its first part looks at two important principles of quality management in education:
- the opening to societal needs
- the importance of stressing the 'act' portion of the PDCA cycle.
The second part of the presentation deals with operational issues in the TQM project.
Exploiting Linked Open Data as Background Knowledge in Data MiningHeiko Paulheim
The document summarizes an approach to exploiting linked open data as background knowledge in data mining tasks. It describes using LOD to generate additional features for machine learning algorithms from entity names in datasets. Experiments show this approach can improve results for classification tasks. Applications discussed include classifying events from Wikipedia and tweets by leveraging background knowledge from DBpedia to prevent overfitting. The document also proposes using LOD to help explain statistics by enriching datasets and analyzing correlations.
Implementing an Integrated Quality Management System in SharePointMontrium
Implementing an Integrated Quality Management
System in SharePoint
For more information on Montrium please visit:
- www.montrium.com
- www.twitter.com/Montrium
- www.youtube.com/Montrium
or email info@montrium.com
Automating Business Processes with SharePointGus Fraser
Making the case for Business Process automation
Workflow options in SharePoint 2010
SharePoint Designer Workflows
Nintex – Workflow for Everyone
Integration and contrast with Microsoft CRM
An overview of SharePoint 2013
Creating workflows with SharePoint Designer 2013 & Visio 2013
The document discusses service quality gaps and how to close them. It identifies four types of gaps: between customer expectations and management perceptions, service design and delivery, delivery and communications, and expectations and performance. For each gap, it provides reasons for the gap and strategies for closing the gap, such as improving marketing research, service standards, employee training, and managing customer expectations.
SharePoint 2013 as a BPM & Workflow Management SystemAndreas Aschauer
SharePoint 2013 as a WFMS
Process Management Services are a crucial piece in any system supporting ECM capabilities. SharePoint 2013 introduced a new technical foundation for workflow management and the modeling of business processes as automated workflows.
In this session the audience will learn the benefits of the new Workflow Management System architecture as well as how to leverage it in real world scenarios. Attendees will see what can be achieved in SharePoint 2013 and Workflow Manager in an easy and controllable fashion. Most important the session will dive into what is not easily feasible Out-of-the-Box and will present a birds-eye view on the BPM ecosystem that has evolved around SharePoint 2010 and 2013.
The key take-away for attendees will be an understanding of how to map real-world process management scenarios onto SharePoint 2013 Workflow workloads and when it is reasonable in terms of project risk and cost of ownership for the SharePoint investment, to move to a third party SharePoint BPM product.
The session is intended for Architects and Decision Makers who already have concrete use cases/scenarios at hand and need to map them to SharePoint technology.
This document provides an overview of SharePoint 2010, including:
- A brief history and evolution of SharePoint products.
- An overview of new features in SharePoint 2010 like the ribbon interface, thin client support, Office Web Apps integration, and social computing features.
- Descriptions of core SharePoint concepts like the server farm, web applications, site collections, sites, lists, and libraries.
- Mentions of tools used to manage and develop solutions for SharePoint like Designer, InfoPath, and Visual Studio.
- Highlights of capabilities like web parts, navigation, theming, workflows, and demonstrations planned for subsequent days.
This document discusses testing and quality assurance for ERP modules. It provides an overview of the testing process roadmap, including establishing requirements and project scope, test planning, case development, different types of testing like unit, integration and user acceptance testing. It also outlines the personnel involved in testing like QA managers, analysts, writers. Metrics for test development and execution are also covered.
Pragmatics Driven Issues in Data and Process Integrity in EnterprisesAmit Sheth
Keynote/Invited Talk
IFIP TC-11 First Working Conference on
Keynote/Invited Talk at the IFIP TC-11 First Working Conference on
Integrity and Internal Control in Information Systems
Zurich, Switzerland, December 4-5, 1997
Here are the answers to the quiz questions:
1. The two main IR system evaluation strategies are:
- System-centered evaluation: Focuses on evaluating different variations of an IR system using document collections, queries, and relevance judgments.
- User-centered evaluation: Focuses on evaluating how well different IR systems satisfy users' actual information needs by testing users on tasks using different systems.
2. For the given information:
- Precision = Relevant docs returned / Total docs returned = 8/18 = 0.44
- Recall = Relevant docs returned / Total relevant docs in collection = 8/20 = 0.4
So the precision is 0.44 and the recall is 0.4
Determination of administrative data quality: recent results and new developm...Piet J.H. Daas
This document outlines a framework developed by Statistics Netherlands to assess the quality of administrative data sources for statistical purposes. It identifies three "hyperdimensions" - source, metadata, and data - to structure 57 quality indicators across 19 dimensions. The framework was tested on 8 administrative data sources using quality checklists. Most sources scored well, though one was found to have delivery and definition issues. The document discusses developing approaches to evaluate the technical and accuracy-related quality of the data dimension. Future work includes fully focusing on the data hyperdimension and studying administrative data quality in a European context through the BLUE-ETS project.
QMS SharePoint Wireframe - download and edit for you useMelissa Jones
This document outlines the proposed site architecture and navigation for a new SharePoint QMS site. It includes wireframes for the site pages, proposed navigation structure with top-level and sub-menus, and assignments for the project team to create lists, libraries, and pages according to the wireframes and then test and deploy the new site.
Part 3 - SharePoint QMS Anyone Can Make - Data DictionaryMelissa Jones
This document provides guidance on setting up lists and libraries in a specific order for a SharePoint quality management system (QMS). It begins by listing standard reference lists that provide quality standards information. It then lists foundational lists for organizing work centers and job descriptions. Core QMS lists and libraries are listed next, including the main document library which can link to other lists. Optional lists for areas like training and customer feedback are also included. Each list and library is then described in more detail regarding its purpose and design.
Jim McCall produced a quality framework model for the US Air Force to bridge the gap between users and developers. The framework defines quality factors divided into software quality, product operation, product revision, and product transition categories. McCall's triangle of quality relates these factors to quality metrics. Product operation factors are defined by metrics expressions to quantify attributes. The approach is user-oriented at the highest level and software-oriented at lower levels, allowing periodic quantification during development.
Metadata Quality Assurance Framework at QQML2016 conference - full versionPéter Király
This document presents a Metadata Quality Assurance Framework to measure and improve metadata quality. It analyzes typical metadata issues like non-informative fields and proposes measuring structural elements like completeness, cardinality, uniqueness, and language specification to predict record quality. Metrics are defined using a problem catalog of known issues mapped to discovery scenarios. Visualizations of early measurement results are shown to identify outliers and inform metadata improvements. The framework is intended to be scalable, transparent, and collaborative.
Quality measurement - How to measure the quality of any object?Grzegorz Grela
THE FRAMEWORK OF QUALITY MEASUREMENT
Quality is the degree to which a set of inherent characteristics fulfils requirements. (ISO 9000)
Requirements and inherent characteristics create finite sets.
Requirements may have both different importance and different values depending on who formulates them.
Requirements do not have to be constant in time.
Quality measurement may be conducted on two levels: analytical and synthetic.
Source: Grela, G. (2015). The Framework of Quality Measurement. Management (18544223), 10(2).
http://www.fm-kp.si/zalozba/ISSN/1854-4231/10_177-191.pdf
15 Months to Certification: Using SharePoint as the Platform for an ISO 9001 ...Barry Peters
Telerx implemented SharePoint to achieve ISO 9001 certification within 15 months. Key aspects included using SharePoint for document control, change control, internal audits, and tracking non-conformances and corrective actions. Workflows automated document approval and status updates. Record control leveraged SharePoint information management policies. The system provided audit trails and version control while supporting continuous improvement processes required by ISO 9001.
The Planning Quality Framework is a collection of tools and techniques that use planning data to help councils understand their development management service performance and benchmark against others. It involves quantitative data like application counts and approval rates, as well as qualitative customer surveys. The framework provides regular reports to give councils insights into the value and quality of their work. It is a low-effort way to focus improvement efforts compared to traditional benchmarking approaches.
The document describes a Quality Management System calibration records list on SharePoint. It provides instructions for accessing calibration records, viewing item details and attachments, running monthly calibration due reports, and exporting the records list to Excel. Maintaining up-to-date calibration records is important for regulatory compliance, and corrective actions may be issued for overdue equipment.
This document summarizes quality assurance and quality control processes. It discusses that quality control focuses on identifying defects in finished products through reactive testing, while quality assurance aims to prevent defects through a proactive process focus. It then outlines the implementation of quality assurance in various areas including stores, production, packing, and batch storage and release. Key activities involve inspection planning, issuing control numbers, monitoring processes, sampling, testing, and record keeping. The goals are to control waste, enhance product quality consistency, and improve customer faith.
The presentation in its first part looks at two important principles of quality management in education:
- the opening to societal needs
- the importance of stressing the 'act' portion of the PDCA cycle.
The second part of the presentation deals with operational issues in the TQM project.
Exploiting Linked Open Data as Background Knowledge in Data MiningHeiko Paulheim
The document summarizes an approach to exploiting linked open data as background knowledge in data mining tasks. It describes using LOD to generate additional features for machine learning algorithms from entity names in datasets. Experiments show this approach can improve results for classification tasks. Applications discussed include classifying events from Wikipedia and tweets by leveraging background knowledge from DBpedia to prevent overfitting. The document also proposes using LOD to help explain statistics by enriching datasets and analyzing correlations.
Implementing an Integrated Quality Management System in SharePointMontrium
Implementing an Integrated Quality Management
System in SharePoint
For more information on Montrium please visit:
- www.montrium.com
- www.twitter.com/Montrium
- www.youtube.com/Montrium
or email info@montrium.com
Automating Business Processes with SharePointGus Fraser
Making the case for Business Process automation
Workflow options in SharePoint 2010
SharePoint Designer Workflows
Nintex – Workflow for Everyone
Integration and contrast with Microsoft CRM
An overview of SharePoint 2013
Creating workflows with SharePoint Designer 2013 & Visio 2013
The document discusses service quality gaps and how to close them. It identifies four types of gaps: between customer expectations and management perceptions, service design and delivery, delivery and communications, and expectations and performance. For each gap, it provides reasons for the gap and strategies for closing the gap, such as improving marketing research, service standards, employee training, and managing customer expectations.
SharePoint 2013 as a BPM & Workflow Management SystemAndreas Aschauer
SharePoint 2013 as a WFMS
Process Management Services are a crucial piece in any system supporting ECM capabilities. SharePoint 2013 introduced a new technical foundation for workflow management and the modeling of business processes as automated workflows.
In this session the audience will learn the benefits of the new Workflow Management System architecture as well as how to leverage it in real world scenarios. Attendees will see what can be achieved in SharePoint 2013 and Workflow Manager in an easy and controllable fashion. Most important the session will dive into what is not easily feasible Out-of-the-Box and will present a birds-eye view on the BPM ecosystem that has evolved around SharePoint 2010 and 2013.
The key take-away for attendees will be an understanding of how to map real-world process management scenarios onto SharePoint 2013 Workflow workloads and when it is reasonable in terms of project risk and cost of ownership for the SharePoint investment, to move to a third party SharePoint BPM product.
The session is intended for Architects and Decision Makers who already have concrete use cases/scenarios at hand and need to map them to SharePoint technology.
This document provides an overview of SharePoint 2010, including:
- A brief history and evolution of SharePoint products.
- An overview of new features in SharePoint 2010 like the ribbon interface, thin client support, Office Web Apps integration, and social computing features.
- Descriptions of core SharePoint concepts like the server farm, web applications, site collections, sites, lists, and libraries.
- Mentions of tools used to manage and develop solutions for SharePoint like Designer, InfoPath, and Visual Studio.
- Highlights of capabilities like web parts, navigation, theming, workflows, and demonstrations planned for subsequent days.
This document discusses testing and quality assurance for ERP modules. It provides an overview of the testing process roadmap, including establishing requirements and project scope, test planning, case development, different types of testing like unit, integration and user acceptance testing. It also outlines the personnel involved in testing like QA managers, analysts, writers. Metrics for test development and execution are also covered.
Pragmatics Driven Issues in Data and Process Integrity in EnterprisesAmit Sheth
Keynote/Invited Talk
IFIP TC-11 First Working Conference on
Keynote/Invited Talk at the IFIP TC-11 First Working Conference on
Integrity and Internal Control in Information Systems
Zurich, Switzerland, December 4-5, 1997
Here are the answers to the quiz questions:
1. The two main IR system evaluation strategies are:
- System-centered evaluation: Focuses on evaluating different variations of an IR system using document collections, queries, and relevance judgments.
- User-centered evaluation: Focuses on evaluating how well different IR systems satisfy users' actual information needs by testing users on tasks using different systems.
2. For the given information:
- Precision = Relevant docs returned / Total docs returned = 8/18 = 0.44
- Recall = Relevant docs returned / Total relevant docs in collection = 8/20 = 0.4
So the precision is 0.44 and the recall is 0.4
Determination of administrative data quality: recent results and new developm...Piet J.H. Daas
This document outlines a framework developed by Statistics Netherlands to assess the quality of administrative data sources for statistical purposes. It identifies three "hyperdimensions" - source, metadata, and data - to structure 57 quality indicators across 19 dimensions. The framework was tested on 8 administrative data sources using quality checklists. Most sources scored well, though one was found to have delivery and definition issues. The document discusses developing approaches to evaluate the technical and accuracy-related quality of the data dimension. Future work includes fully focusing on the data hyperdimension and studying administrative data quality in a European context through the BLUE-ETS project.
Analytics and reporting context linkedin finalDennis Crow
This document discusses enterprise information architecture for analytics and reporting. It provides definitions and principles for information architecture, including that it synthesizes analytical requirements and data capabilities, and that information is the outcome of how users interpret data. The document outlines relationships between analytics stakeholders and describes the data warehousing, analytics, and performance measurement process. It also includes diagrams showing an analysis process and how information can be presented from different data sources.
Sharon Dawes (CTG Albany) Open data quality: a practical viewOpen City Foundation
This document discusses open data quality and focuses on ensuring data is fit for its intended use. It notes that while open data aims to provide easy access, the value depends on the quality and how users apply the data. Quality issues can arise from how data is originally collected and maintained by different government systems. The document recommends open data providers adopt stewardship practices to maintain metadata and ensure quality, while users should approach data cautiously and look for ways to engage in data communities. Overall it promotes openness but also a realistic view of potential quality problems and the need for tools and strategies to maximize data value for various users.
Semantic Similarity and Selection of Resources Published According to Linked ...Riccardo Albertoni
The position paper aims at discussing the potential of exploiting linked data best practice to provide metadata documenting domain specific resources created through verbose acquisition-processing pipelines. It argues that resource selection, namely the process engaged to choose a set of resources suitable for a given analysis/design purpose, must be supported by a deep comparison of their metadata. The semantic similarity proposed in our previous works is discussed for this purpose and the main issues to make it scale up to the web of data are introduced. Discussed issues contribute beyond the re-engineering of our similarity since they largely apply to every tool which is going to exploit information made available as linked data. A research plan and an exploratory phase facing the presented issues are described remarking the lessons we have learnt so far.
Data Mining is the process of discovering new correlations, patterns, and trends by digging into (mining) large amounts of data stored in warehouses, using artificial intelligence, statistical and mathematical techniques. Data mining can also be defined as the process of extracting knowledge hidden from large volumes of raw data i.e. the nontrivial extraction of implicit, previously unknown, and potentially useful information from data. The alternative name of Data Mining is Knowledge discovery (mining) in databases (KDD), knowledge extraction, data/pattern analysis, etc.
Towards Privacy-Preserving Evaluation for Information Retrieval Models over I...Twitter Inc.
This document presents a framework called Privacy-Preserving Evaluation (PPE) that allows researchers to evaluate information retrieval models on private industrial datasets without directly accessing the raw data. PPE is designed as a level 2 evaluation platform that allows users to explore indexes, inspect results, and implement retrieval models through APIs. Experiments show that several retrieval models perform well on an industrial collection under PPE, and diagnostic analyses involving document perturbations provide insights into model behaviors without revealing private data. The level-based PPE framework enables more effective use of valuable private datasets while preserving privacy.
This document presents a framework for evaluating health IT projects. It consists of several components: [1] A project structure template to guide planning, preparation, evaluation, and dissemination of results. [2] Multidimensional evaluation methods that assess both qualitative and quantitative outcomes across technical, clinical, and organizational areas. [3] Criteria pools for selecting evaluation measures. [4] Guidelines for confidentiality, analysis, and reporting of results. The goal is to provide consistent, high-quality evaluation that identifies benefits and areas for improvement to inform future health IT implementations.
Changing the Curation Equation: A Data Lifecycle Approach to Lowering Costs a...SEAD
This document discusses the Sustainable Environment Actionable Data (SEAD) project, which aims to lower the costs and increase the value of data curation through a data lifecycle approach. SEAD provides lightweight data services to support sustainability research, including secure project workspaces, active and social curation tools, and integrated lifecycle support for data from ingest to long-term preservation. By leveraging technologies like Web 2.0 and standards, SEAD simplifies and automates curation processes using metadata captured from data producers and users. This allows curation activities to begin earlier in the data lifecycle and be distributed across researchers and curators.
This document discusses research methodology, including research, data collection methods, and measurement scales. It defines research as the systematic process of searching for information on a predetermined topic. The main types of research are descriptive vs analytical, applied vs fundamental, quantitative vs qualitative, and conceptual vs empirical. Primary and secondary data collection methods are outlined, including questionnaires, interviews, observation, and schedules for primary data, as well as internal and external sources for secondary data. Finally, the key measurement scales are defined as nominal, ordinal, interval, and ratio scales. Self-rating scales like Likert scales are also introduced.
Big Data Journeys: Review of roadmaps taken by early adopters to achieve thei...Krishnan Parasuraman
Implementing a Big Data program can be a long and arduous journey. Each organization has its own unique business drivers and technical considerations that drive their big data adoption roadmaps. Whatever be your organization's specific big data driver - be it managing a rapid surge of data, implementing a new set of analytic capabilities, incorporating unstructured data as part of your enterprise data platform or accessing real time information for actionable intelligence - the approach and roadmap that you put in place to reach that end goal becomes all the more critical in a space where early success stories are relatively rare, skill sets are hard to find and technologies are still evolving.
In this session we will chronicle the journeys of four different organizations that were early adopters of big data. Each of them charted a different path to achieve their big data goals. We will look at what were the key drivers behind their respective approaches, what worked and what did not work for them.
The document discusses quality evaluation of geospatial data. It summarizes the key successes of work package 8, which include utilizing open standards to define quality, developing a common understanding of what quality means for specifications and user requirements, and how to measure it. Automating quality evaluation services and providing quality results as metadata are also highlighted. Benefits of quality evaluation include early error detection, reduced costs, consistent procedures, improved analysis, and trusted data. The ESDIN approach to quality involves developing quality models, rulesets, templates and an object-oriented geospatial rules engine to automate quality evaluation.
The document discusses developing indicators for monitoring forest governance. It outlines a workshop that will cover creating reliable indicators, IFRI research and data, preliminary analyses using random forests, and how indicators can inform interventions. IFRI has long-term data on social, ecological and institutional factors from sites in 11 countries. Random forest analysis is useful for exploring these complex datasets and identifying important variables and interactions that predict outcomes of interest for forest governance. The results can indicate what actions may positively or negatively impact forest outcomes depending on local conditions and causal processes. Developing reliable indicators is an ongoing process that adapts to changing relationships and behaviors over time and place.
This document discusses data management for the CGIAR Research Program on Dryland Systems. It provides an overview of ICARDA's current status on data management, key sources of data under the CRP Dryland Systems, and issues related to research and data quality. The document proposes a workflow for improving data management and sharing that involves validating data, archiving it, and establishing permissions and approvals for sharing. It emphasizes training partners, identifying risks, and monitoring impact to enhance data quality and open access for CRP Dryland Systems research.
Prof. Melinda Laituri, Colorado State University | Map Data Integrity | SotM ...Kathmandu Living Labs
State of the Map Asia (SotM-Asia) is the annual regional conference of OpenStreetMap (OSM) organized by OSM communities in Asia. First SotM-Asia was organized in Jakarta, Indonesia in 2015, and the second was organized in Manila, Philippines in 2016. This year’s conference, third in the series, was organized in Kathmandu, Nepal on September 23 – 24, 2017 at Park Village Resort, Budhanilkantha, Kathmandu, Nepal.
We brought nearly 200 Open Mapping enthusiasts from Asia and beyond to this year’s SotM-Asia. The event provided an opportunity to share knowledge and experience among mappers; expand their network; and generate ideas to expand map coverage and effective use of OSM data in Asian continent. We chose ‘from creation to use of OSM data’ as the theme of this year’s conference, emphasizing on the effective use of OSM data. We also brought together a government panel from four different countries in this year’s SotM-Asia. We believe this event will deepen the bond and enhance collaboration among OSM communities across Asia.
More information about the conference can be found on: http://stateofthemap.asia.
This document presents a draft maturity matrix for long-term scientific data stewardship. The matrix defines 5 levels of maturity for 10 key components of data stewardship, including preservation, accessibility, usability, production sustainability, and data quality. Each increasing level represents more advanced and formalized approaches to managing the data according to established standards and community best practices. The authors thank various subject matter experts who helped define the maturity levels based on their expertise in areas such as data archiving, access, and product development.
This document provides an overview of key concepts related to data and big data. It defines data, digital data, and the different types of digital data including unstructured, semi-structured, and structured data. Big data is introduced as the collection of large and complex data sets that are difficult to process using traditional tools. The importance of big data is discussed along with common sources of data and characteristics. Popular tools and technologies for storing, analyzing, and visualizing big data are also outlined.
Similar to Proposal for a quality framework for the evaluation of administrative and survey data (20)
Big Data and official statistics with examples of their usePiet J.H. Daas
The document provides an overview of the work done by the Center for Big Data Statistics (CBDS) at Statistics Netherlands. It discusses several examples of using big data sources to produce official statistics:
1) Road sensor data was used to produce the first official big data-based statistics on traffic intensity and its correlation with GDP.
2) Mobile phone data was analyzed to produce statistics on mobility patterns, daytime population, and tourism.
3) AIS ship tracking data was analyzed to study ship movements and transhipment locations.
4) Web scraping and text analysis of company websites was used to identify innovative companies, including small companies not covered by traditional surveys.
5) Sentiment analysis and
ESSnet Big Data WP8 Methodology (+ Quality, +IT)Piet J.H. Daas
1. The documents discuss methodology, quality, and IT aspects of big data within the ESSnet Big Data project.
2. Key topics addressed include the big data processing lifecycle, metadata management challenges, and quality aspects like coverage, accuracy, and comparability over time.
3. Common themes that emerged across work packages include the need for a unified framework for data integration and metadata, and the value of shared software and training resources.
Piet Daas and Marco Puts from Statistics Netherlands presented on big data methods and techniques. They discussed the four phases of working with big data: collect, process, analyze, and disseminate. They provided examples of each phase using road sensor data to measure traffic, scraping company websites to identify innovative firms, and using aerial images to detect solar panels. They emphasized the need to preprocess and clean big data due to its noisy nature. When analyzing big data, they discussed dealing with imbalanced datasets, such as through oversampling rare cases. They concluded by showing examples of visualizing big data results as dot maps and animations.
Isi 2017 presentation on Big Data and biasPiet J.H. Daas
1) The document discusses three types of using big data in statistics: (1) combined with survey data, (2) from a single complete source, and (3) from a single incomplete source.
2) Examples of type 2 include road sensor traffic data and web-scraped price data. These sources completely cover their target populations.
3) Examples of type 3 include social media data and mobile phone data. Only part of the target population is included, so ways must be found to deal with the missing part, such as determining the characteristics of the included population.
Responsible Data Science at Statistics NetherlandsPiet J.H. Daas
Piet Daas presents on responsible data science at Statistics Netherlands and implications for big data research. Some key points:
- Statistics Netherlands uses a variety of administrative data and surveys in its Social Statistical Database to produce statistics, ensuring privacy through anonymization and access restrictions.
- The Center for Big Data Statistics aims to produce new real-time statistics using big data sources while reducing data collection burdens and advancing methodology. Challenges include a lack of established big data methods and ensuring transparency of models.
- Responsible data science principles of fairness, accuracy, confidentiality and transparency must be further developed to fully leverage big data's potential while preventing harms, such as through de-identification and model explainability.
CBS lecture at the opening of Data Science Campus of ONSPiet J.H. Daas
The document summarizes work done at the Center for Big Data Statistics, including case studies and methodological research. Some examples of projects are:
1) Visualizing income data in 2D and 3D heat maps showing relationships between age, income, and amount.
2) Analyzing road sensor data to show relationships between traffic intensity and GDP.
3) Tracking "ginger bread" product sales from scanner data around Saint Nicolas festivities.
4) Developing a social tension indicator using Twitter data.
5) Identifying web-only shops and innovative companies using web page archives.
1) Statistics Netherlands is working on several Big Data projects to produce new official statistics in a timely manner using large alternative data sources like road sensors.
2) Their Center for Big Data Statistics aims to reduce response burden, deepen methodological knowledge, and stimulate cooperation using an ecosystem of partners.
3) As a proof of concept, they have produced the first Big Data-based official statistic on regional traffic intensity using minute-level road sensor data from 20,000 sensors on Dutch highways. This required data cleaning, transformation, estimation techniques, and integrating skills from statistics, IT, and subject-matter expertise.
Extracting information from ' messy' social media dataPiet J.H. Daas
This document summarizes research conducted by Statistics Netherlands on using social media data for official statistics. It discusses (1) determining sentiment in social media and its correlation with consumer confidence surveys, (2) developing an initial social media-based (un)safety monitor, and (3) analyzing the composition of social media users to determine background characteristics like gender. Key findings include high correlation between sentiment and surveys, the ability to predict consumer confidence from sentiment, and developing accurate methods to classify user gender using multiple social media signals.
This document summarizes the experiences of Statistics Netherlands with big data research. It discusses two types of data - primary data collected through surveys and secondary data from administrative sources and big data. It provides examples of big data research conducted using road sensor data, mobile phone data, and social media data. Lessons learned include the need for skills in accessing and analyzing large datasets, dealing with noisy unstructured data, and addressing privacy and costs. Important future research topics mentioned are profiling units in big data, data editing at large scale, and data reduction techniques.
Profiling Big Data sources to assess their selectivityPiet J.H. Daas
This document discusses profiling Big Data sources to assess their selectivity. It analyzes a random sample of 1,000 Dutch Twitter users to determine gender selectivity. Several methods are used to infer gender from profile elements: (1) First names are analyzed using a Dutch name database, (2) Bios and tweets are examined for gendered language, (3) Pictures are processed with face recognition software. Overall results show first names provided the highest diagnostic odds ratio for determining gender, while profile pictures provided the lowest. The study aims to develop clever ways to combine these methods for more accurate gender profiling of social media users.
Using Road Sensor Data for Official Statistics: towards a Big Data MethodologyPiet J.H. Daas
This document discusses using road sensor data for official statistics in the Netherlands. It describes challenges around dealing with large volumes of data, creating historical time series, and ensuring accuracy. A statistical process is outlined that cleans, transforms, selects, estimates from and frames the raw road sensor data, which records over 230 million vehicle counts per day. Key steps include selecting only necessary variables from valid data on main routes, putting daily records together, cleaning using recursive Bayesian estimation and a hidden Markov model, and estimating traffic indices from the cleaned data.
Big Data @ CBS for Fontys students in EindhovenPiet J.H. Daas
This document summarizes the experiences of Statistics Netherlands with big data. It discusses two types of data - primary data from their own surveys and secondary data from other sources like administrative records and big data. It provides examples of exploratory big data studies conducted using road sensor data, mobile phone data, and social media data. It finds that combining IT skills with statistical methodology is important for working with big data. Skills in data science, machine learning, and extracting information from diverse sources like text and images are needed. The document also discusses lessons learned regarding the types of big data, accessing and analyzing large volumes of data, dealing with noisy and unstructured data, and moving beyond simple correlation.
Quality challenges in modernising business statisticsPiet J.H. Daas
This document discusses quality challenges in modernizing official business statistics due to two fundamental changes: commercialization of statistics and globalization. It notes these changes have resulted in different statistics being needed and different ways of producing statistics, using available administrative data, secondary sources, and modernizing survey methodology. It evaluates the qualities and disqualities of various data sources like surveys, administrative data, and big data. Finally, it tentatively concludes statistical systems need to develop new indicators and integrate multiple data sources using their individual qualities, while also modernizing business surveys. The goal is to update multi-source and mixed-mode strategies for producing official statistics.
This document discusses quality approaches for big data in statistics. It outlines limitations of established quality frameworks for big data, including population not being known, unbalanced data coverage, and unclear relevance of data sources. Options presented to address these limitations include deriving background information, using modeling approaches, and calibration or correlation studies. The document advocates that statistical organizations validate information from other big data producers, get to know big data sources, use big data for efficiency and early indicators, and create an environment conducive to innovative big data approaches.
Temple of Asclepius in Thrace. Excavation resultsKrassimira Luka
The temple and the sanctuary around were dedicated to Asklepios Zmidrenus. This name has been known since 1875 when an inscription dedicated to him was discovered in Rome. The inscription is dated in 227 AD and was left by soldiers originating from the city of Philippopolis (modern Plovdiv).
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxEduSkills OECD
Iván Bornacelly, Policy Analyst at the OECD Centre for Skills, OECD, presents at the webinar 'Tackling job market gaps with a skills-first approach' on 12 June 2024
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.pptHenry Hollis
The History of NZ 1870-1900.
Making of a Nation.
From the NZ Wars to Liberals,
Richard Seddon, George Grey,
Social Laboratory, New Zealand,
Confiscations, Kotahitanga, Kingitanga, Parliament, Suffrage, Repudiation, Economic Change, Agriculture, Gold Mining, Timber, Flax, Sheep, Dairying,
Chapter wise All Notes of First year Basic Civil Engineering.pptxDenish Jangid
Chapter wise All Notes of First year Basic Civil Engineering
Syllabus
Chapter-1
Introduction to objective, scope and outcome the subject
Chapter 2
Introduction: Scope and Specialization of Civil Engineering, Role of civil Engineer in Society, Impact of infrastructural development on economy of country.
Chapter 3
Surveying: Object Principles & Types of Surveying; Site Plans, Plans & Maps; Scales & Unit of different Measurements.
Linear Measurements: Instruments used. Linear Measurement by Tape, Ranging out Survey Lines and overcoming Obstructions; Measurements on sloping ground; Tape corrections, conventional symbols. Angular Measurements: Instruments used; Introduction to Compass Surveying, Bearings and Longitude & Latitude of a Line, Introduction to total station.
Levelling: Instrument used Object of levelling, Methods of levelling in brief, and Contour maps.
Chapter 4
Buildings: Selection of site for Buildings, Layout of Building Plan, Types of buildings, Plinth area, carpet area, floor space index, Introduction to building byelaws, concept of sun light & ventilation. Components of Buildings & their functions, Basic concept of R.C.C., Introduction to types of foundation
Chapter 5
Transportation: Introduction to Transportation Engineering; Traffic and Road Safety: Types and Characteristics of Various Modes of Transportation; Various Road Traffic Signs, Causes of Accidents and Road Safety Measures.
Chapter 6
Environmental Engineering: Environmental Pollution, Environmental Acts and Regulations, Functional Concepts of Ecology, Basics of Species, Biodiversity, Ecosystem, Hydrological Cycle; Chemical Cycles: Carbon, Nitrogen & Phosphorus; Energy Flow in Ecosystems.
Water Pollution: Water Quality standards, Introduction to Treatment & Disposal of Waste Water. Reuse and Saving of Water, Rain Water Harvesting. Solid Waste Management: Classification of Solid Waste, Collection, Transportation and Disposal of Solid. Recycling of Solid Waste: Energy Recovery, Sanitary Landfill, On-Site Sanitation. Air & Noise Pollution: Primary and Secondary air pollutants, Harmful effects of Air Pollution, Control of Air Pollution. . Noise Pollution Harmful Effects of noise pollution, control of noise pollution, Global warming & Climate Change, Ozone depletion, Greenhouse effect
Text Books:
1. Palancharmy, Basic Civil Engineering, McGraw Hill publishers.
2. Satheesh Gopi, Basic Civil Engineering, Pearson Publishers.
3. Ketki Rangwala Dalal, Essentials of Civil Engineering, Charotar Publishing House.
4. BCP, Surveying volume 1
A Visual Guide to 1 Samuel | A Tale of Two HeartsSteve Thomason
These slides walk through the story of 1 Samuel. Samuel is the last judge of Israel. The people reject God and want a king. Saul is anointed as the first king, but he is not a good king. David, the shepherd boy is anointed and Saul is envious of him. David shows honor while Saul continues to self destruct.
THE SACRIFICE HOW PRO-PALESTINE PROTESTS STUDENTS ARE SACRIFICING TO CHANGE T...indexPub
The recent surge in pro-Palestine student activism has prompted significant responses from universities, ranging from negotiations and divestment commitments to increased transparency about investments in companies supporting the war on Gaza. This activism has led to the cessation of student encampments but also highlighted the substantial sacrifices made by students, including academic disruptions and personal risks. The primary drivers of these protests are poor university administration, lack of transparency, and inadequate communication between officials and students. This study examines the profound emotional, psychological, and professional impacts on students engaged in pro-Palestine protests, focusing on Generation Z's (Gen-Z) activism dynamics. This paper explores the significant sacrifices made by these students and even the professors supporting the pro-Palestine movement, with a focus on recent global movements. Through an in-depth analysis of printed and electronic media, the study examines the impacts of these sacrifices on the academic and personal lives of those involved. The paper highlights examples from various universities, demonstrating student activism's long-term and short-term effects, including disciplinary actions, social backlash, and career implications. The researchers also explore the broader implications of student sacrifices. The findings reveal that these sacrifices are driven by a profound commitment to justice and human rights, and are influenced by the increasing availability of information, peer interactions, and personal convictions. The study also discusses the broader implications of this activism, comparing it to historical precedents and assessing its potential to influence policy and public opinion. The emotional and psychological toll on student activists is significant, but their sense of purpose and community support mitigates some of these challenges. However, the researchers call for acknowledging the broader Impact of these sacrifices on the future global movement of FreePalestine.
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...TechSoup
Whether you're new to SEO or looking to refine your existing strategies, this webinar will provide you with actionable insights and practical tips to elevate your nonprofit's online presence.
Proposal for a quality framework for the evaluation of administrative and survey data
1. A quality framework
for the evaluation of
administrative and survey data
Piet J.H. Daas, Judit Arends-Tóth,
Barry Schouten and Léander Kuivenhoven
Statistics Netherlands
2. Overview
Reason for work
View on Quality
Starting point
Literature study results
Overview of the framework
Application
Future work
3. Reason for work
Statistics Netherlands increases the use of
data (sources) collected and maintained by
others
• To decrease response burden and costs
As a result:
• More dependent on external data sources
• Must be able to monitor the quality of external
data sources
– Develop a way to monitor the quality of external
data sources
4. View on quality
Statistics Netherlands definition of the quality of
external data sources:
“Usability for the production of statistics”
Differs from quality as used by the data source
maintainer
5. Starting point
Previous work at Statistics Netherlands
• Most recent: Work of Daas and Fonville
– Determination of register quality
– Described in a paper for the register seminar in Finland
(2007)
• Improvement is possible
– Predominantly business register data
– International experience could be included more
New project started
• Develop quality framework for the evaluation of
external data sources (focus on registers and
administrative data)
6. Literature study (1)
Performed and extensive literature study
• Publically available papers/books that studied
the quality of administrative data sources and
registers → Dutch and English
• A lot of research focuses on quality of survey
collected → these were excluded
• Ended up with quite a limited lists of important
papers:
– Book of the Wallgren’s (S)
– Daas and Fonville (NL)
– Eurostat paper on Quality of administrative Data
– Work performed at ONS and by Thomas (UK)
– UNECE paper of Nordic countries
7. Literature study (2)
Conclusions:
• A general level of mutuality
– The papers identified many similar quality aspects
(quality indicators)
• None of the ‘views’ on quality were exactly alike
• How to combine all these views?
• Something higher than a dimension was needed
• Karr et al. 2006 used the term Hyperdimension to
distinguish different views on quality
• Combine all quality aspects identified in all studies
and new aspects in a single framework !!
8. Quality framework
Framework has 4 hyperdimensions
• Four views on the quality of the external data source
The hyperdimensions identified are:
• Source → Data source as a whole
• Metadata → Conceptual metadata of data in source
• Data → Facts (values) in data source
• Process → Processing related quality aspects
9. Quality framework levels
Levels distinguished:
HYPERDIMENSION
n>1
DIMENSION
n >= 1
QUALITY INDICATOR
1:n
Measurement method
10. 1) Source hyperdimension
Here the data source is viewed upon as a file
delivered by the data source maintainer to
the NSI
Dimensions (5):
• Supplier, Relevance, Privacy and security,
Delivery, and Procedures
11. Source hyperdimension
Hyper- Dimension Indicator Measurement method
dimension
Source 1. Supplier
Supplier Contact Name, Contact information
1.1 Contact - Name of the data source
- Data source contact info
Relevance Adm. burden Effect of use on adm. burden
- NSI a contact person
of NSI
1.2 Purpose (time and money)
- Reason for use of the data
source by NSI
Privacy and Legal provision Check if Personal Data
2. Relevance
security2.1 Usefulness - Importance data source for NSI
Protection act applies
Delivery2.2 Envisaged use
Costs - Potentialuse for NSI use of data
Costs of statistical
source
2.3 Information - Does the data source satisfy
12. 2) Metadata hyperdimension
Focuses on the conceptual metadata quality
aspects of the data source.
Other metadata aspects (such as process
meta) are not included
Dimensions (4):
• Clarity, Comparability, Unique keys, and Data
treatment by data source maintainer
13. Metadata hyperdimension
Hyper- Dimension Indicator Measurement method
dimension
Metadata Clarity
1. Clarity Population Description of the population
1.1 Population definition - Clarity scoredata source
definition used in of the definition
Unique Definition of variables
1.2 keys Identification - Clarity scoreunique keys
Presence of of the definition
(and categories)
keys present (which)
1.3 Time dimensions - Clarity score of the definition
Data 1.4 Geographic demarcation - Clarity score of the definition
treatment Checks Variable value checks
by data source performed
1.5 Definition changes
maintainer - Familiarity with occurred
changes
2. Comparability Modifications Familiarity with data
modifications
2.1 Population definition - Comparability with NSI
comparison definition
14. 3) Data hyperdimension
Aspects related to data in the data source
• All aspects are accuracy related
Actively being discussed at our office
• Future changes may be possible
Dimensions (9)
• Over coverage, Under coverage, Linkability, Unit
non-response, Item non-response, Measurement,
Processing, Precision, and Sensitivity
Remark: Precision was mainly included for (externally collected) survey data
15. Data hyperdimension
Hyper- Dimension Indicator Measurement method
dimension
Data Over coverage Non-pop. units
1. Over coverage Percentage of units not 1.1
Non-population units - Percentage of units to population of NSI
belonging not
belonging to population
Linkability Linkable units Percentage if units linked
2. Under coverage
2.1 Missing units - Percentage of missing
Measurement Incompatible population units with violated
Fraction of fields
records edit rules
2.2 Selectivity - R-index for population
Processing Adjustment composition
Fraction of fields adjusted
2.3 Effect on core variables - Maximum bias of average for
Imputation Fraction of fields imputed
core variable
- Maximum RMSE of average
R-index: Representative index; RMSE: Root mean square Error; MSE: Mean Square Error
16. 4) Process hyperdimension
Focuses on the processing of the data source
• by the data source maintainer
• by the NSI
Not discussed here, future work
Framework was developed without specifically
focusing at process related quality aspects
• main focus was product related
17. Framework and external data sources
Developed for administrative data
• Registers a.o.
Why not use it for surveys?
• In the case were the data is collected by an
organization other than Statistics Netherlands
• Experiences in the past resulted in many quality
related discussions
– transparency of data collection process?
Resulted in some minor adjustments of the
framework
• Some terms were adjusted
• Review with sample approach in mind
18. Application of the framework
Should be applied to externally collected data
sources
• Administrative data, registers, non-NSI surveys
How to apply?
• Source and Metadata hyperdimension
– Checklists have been developed
• Data hyperdimension
– Methods of calculation have been proposed
– Currently looking at a practical means to apply these
• Process hyperdimension
– Under investigation
19. Application of the framework (2)
For each data source and use of data source
1) Evaluate Source with checklist
– 2 ways: a quick an a complete scan
– When no problems occur continue
2) Evaluate Metadata with checklist
– 3 ways: replace, additional or new
– When no problems occur continue
3) Evaluate Data
– In a standardized way (scripts or computer program)
– Probably requires some very specific test at the end
– Framework should be generally applicable
– User mostly has a specific use in mind
20. Future work
Evaluate registers and external survey data
• Is the framework generally applicable for all
sources?
Thoroughly test Source and Metadata checklists
• Feed-back on usability by users
Calculation methods for Data
• A single way of determining every Quality indicator
Study how to efficiently evaluate Data
• E.g. Scripts or computer program
Determine the quality aspects in the Process
hyperdimension