Machine Learning encompasses data acquisition, transmission, retention, analysis, and reduction. The expected outgrowth of 24x7 data systems and operations centers is Knowledge Engineering and Data Intensive Analytics AKA Machine Learning. This presentation will develop and apply Machine Learning concepts to the Upstream O&G industry. Specific focus will be given to the fundamental concepts and definitions of Machine Learning along with the application of Machine Learning.
E&P organizations are turning more attention to accumulated data to enhance operating efficiency, safety, and recovery. The computing paradigm is shifting, the O&G paradigm is shifting, and the rise of the machine learning paradigm requires careful attention to top-down integrated systems engineering. A system approach will be presented to stimulate out-of-the-box thinking to address the machine learning paradigm.
Data for Impact - Horizon 2020 project pioneering big data approaches for improved assessment of the societal impact in the health, demographic change and well-being societal challenge at national and EU levels. Data4Impact aspires to develop a set of new indicators for assessing research and innovation performance based on a hands-on and data-driven approach.
Here is the presentation from the Data4Impact workshop, which took place on 24th of September 2018.
Efficient O&G does not suffice in an industry downturn – effective investment in time and effort is required to rise above the pack
Production analysis need not be mystical; it should not be rote
Nuance and subtle variations provide leading indicators into impending production issues
Decline curves, certainly crucial, must be analyzed in context
Case-based, topological analysis, rule inference, curve plotting solutions are common solutions, but fall short
Application of nuance analysis within environment of Data-Intensive Scientific Discovery
Machine Learning encompasses data acquisition, transmission, retention, analysis, and reduction. The expected outgrowth of 24x7 data systems and operations centers is Knowledge Engineering and Data Intensive Analytics AKA Machine Learning. This presentation will develop and apply Machine Learning concepts to the Upstream O&G industry. Specific focus will be given to the fundamental concepts and definitions of Machine Learning along with the application of Machine Learning.
E&P organizations are turning more attention to accumulated data to enhance operating efficiency, safety, and recovery. The computing paradigm is shifting, the O&G paradigm is shifting, and the rise of the machine learning paradigm requires careful attention to top-down integrated systems engineering. A system approach will be presented to stimulate out-of-the-box thinking to address the machine learning paradigm.
Data for Impact - Horizon 2020 project pioneering big data approaches for improved assessment of the societal impact in the health, demographic change and well-being societal challenge at national and EU levels. Data4Impact aspires to develop a set of new indicators for assessing research and innovation performance based on a hands-on and data-driven approach.
Here is the presentation from the Data4Impact workshop, which took place on 24th of September 2018.
Efficient O&G does not suffice in an industry downturn – effective investment in time and effort is required to rise above the pack
Production analysis need not be mystical; it should not be rote
Nuance and subtle variations provide leading indicators into impending production issues
Decline curves, certainly crucial, must be analyzed in context
Case-based, topological analysis, rule inference, curve plotting solutions are common solutions, but fall short
Application of nuance analysis within environment of Data-Intensive Scientific Discovery
HOBBIT project overview presented at European Big Data Value Forum, 21-23 Nov 2017, held in Versailles, France (Palais des Congres).
This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227).
Selection of Articles Using Data Analytics for Behavioral Dissertation Resear...PhD Assistance
Outcomes in health-related issues including psychological, educational, Behavioral, environmental, and social are intended to sustain positive change by digital interferences. These changes may be delivered using any digital device like a phone or computer, and make them gainful for the provider. Complex and large-scale datasets that contain usage data can be yielded by testing a digital intervention. This data provides invaluable detail about how the users interact with these interventions and notify their knowledge of engagement, if they are analyzed properly. This paper recommends an innovative framework for the process of analyzing usage associated with a digital intervention .
PhD Assistance is an Academic The Best Dissertation Writing Service & Consulting Support Company established in 2001. specialiWeze in providing PhD Assignments, PhD Dissertation Writing Help , Statistical Analyses, and Programming Services to students in the USA, UK, Canada, UAE, Australia, New Zealand, Singapore and many more.
Website Visit: https://bit.ly/3dANXUD
Contact Us:
UK NO: +44-1143520021
India No: +91-8754446690
Email: info@phdassistance.com
Presentation about new data, methods and outputs to create knowledge for innovation policy. Presented at the OECD Blue Sky Conference, 20 September 2016.
It is the basics of what is Big Data and What are the challenges for Big Data Analysis. This slides are for beginners of computer engineering....Hope this will be somewhat helpful!!!
Big data impact on society: a research roadmap for Europe (BYTE project resea...Anna Fensel
With its rapid growth and increasing adoption, big data is producing a growing impact in society. Its usage is opening both opportunities such as new business models and economic gains and risks such as privacy violations and discrimination. Europe is in need of a comprehensive strategy to optimise the use of data for a societal benefit and increase the innovation and competitiveness of its productive activities. In this paper, we contribute to the definition of this strategy with a research roadmap that considers the positive and negative externalities associated with big data, maps research and innovation topics in the areas of data management, processing, analytics, protection, visualisation, as well as non-technical topics, to the externalities they can tackle, and provides a time frame to address these topics.
Data Mining of Project Management Data: An Analysis of Applied Research Studies.Gurdal Ertek
Data collected and generated through and posterior to projects, such as data residing in project management software and post project review documents, can be a major source of actionable insights and competitive advantage. This paper presents a rigorous
methodological analysis of the applied research published in academic literature, on the application of data mining (DM) for project management (PM). The objective of the paper is to provide a comprehensive analysis and discussion of where and how data mining is applied for project management data and to provide practical insights for future research in the field.
https://dl.acm.org/citation.cfm?id=3176714
https://ertekprojects.com/ftp/papers/2017/ertek_et_al_2017_Data_Mining_of_Project_Management_Data.pdf
Patterns for Successful Data Science Projects (Spark AI Summit)Bill Chambers
Running data science workloads is challenge regardless of whether you are running them on your laptop, on an on-premises cluster, or in the cloud. While buying 100% managed service is an option, these tools can be expensive and lack extensibility. Therefore, many companies option for open source data science tools like scikit-learn and Apache Spark’s MLlib in order to balance both functionality and cost.
However, even if a project succeeds at a point in time with any set of tools, these projects become harder and harder to maintain as data volumes increase and a desire for real-time pushes technology to its limit. New projects also struggle as new challenges of scale invalidate previous assumptions.
This talk will discuss some patterns that we see at Databricks that companies leverage to succeed with their data science projects. Key takeaways will be:
– Striving for simplicity
– Removing cognitive load for you and your team
– Working with data, big and small
– Effectively leveraging the ecosystem of tools to be successful
HOBBIT project overview presented at European Big Data Value Forum, 21-23 Nov 2017, held in Versailles, France (Palais des Congres).
This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227).
Selection of Articles Using Data Analytics for Behavioral Dissertation Resear...PhD Assistance
Outcomes in health-related issues including psychological, educational, Behavioral, environmental, and social are intended to sustain positive change by digital interferences. These changes may be delivered using any digital device like a phone or computer, and make them gainful for the provider. Complex and large-scale datasets that contain usage data can be yielded by testing a digital intervention. This data provides invaluable detail about how the users interact with these interventions and notify their knowledge of engagement, if they are analyzed properly. This paper recommends an innovative framework for the process of analyzing usage associated with a digital intervention .
PhD Assistance is an Academic The Best Dissertation Writing Service & Consulting Support Company established in 2001. specialiWeze in providing PhD Assignments, PhD Dissertation Writing Help , Statistical Analyses, and Programming Services to students in the USA, UK, Canada, UAE, Australia, New Zealand, Singapore and many more.
Website Visit: https://bit.ly/3dANXUD
Contact Us:
UK NO: +44-1143520021
India No: +91-8754446690
Email: info@phdassistance.com
Presentation about new data, methods and outputs to create knowledge for innovation policy. Presented at the OECD Blue Sky Conference, 20 September 2016.
It is the basics of what is Big Data and What are the challenges for Big Data Analysis. This slides are for beginners of computer engineering....Hope this will be somewhat helpful!!!
Big data impact on society: a research roadmap for Europe (BYTE project resea...Anna Fensel
With its rapid growth and increasing adoption, big data is producing a growing impact in society. Its usage is opening both opportunities such as new business models and economic gains and risks such as privacy violations and discrimination. Europe is in need of a comprehensive strategy to optimise the use of data for a societal benefit and increase the innovation and competitiveness of its productive activities. In this paper, we contribute to the definition of this strategy with a research roadmap that considers the positive and negative externalities associated with big data, maps research and innovation topics in the areas of data management, processing, analytics, protection, visualisation, as well as non-technical topics, to the externalities they can tackle, and provides a time frame to address these topics.
Data Mining of Project Management Data: An Analysis of Applied Research Studies.Gurdal Ertek
Data collected and generated through and posterior to projects, such as data residing in project management software and post project review documents, can be a major source of actionable insights and competitive advantage. This paper presents a rigorous
methodological analysis of the applied research published in academic literature, on the application of data mining (DM) for project management (PM). The objective of the paper is to provide a comprehensive analysis and discussion of where and how data mining is applied for project management data and to provide practical insights for future research in the field.
https://dl.acm.org/citation.cfm?id=3176714
https://ertekprojects.com/ftp/papers/2017/ertek_et_al_2017_Data_Mining_of_Project_Management_Data.pdf
Patterns for Successful Data Science Projects (Spark AI Summit)Bill Chambers
Running data science workloads is challenge regardless of whether you are running them on your laptop, on an on-premises cluster, or in the cloud. While buying 100% managed service is an option, these tools can be expensive and lack extensibility. Therefore, many companies option for open source data science tools like scikit-learn and Apache Spark’s MLlib in order to balance both functionality and cost.
However, even if a project succeeds at a point in time with any set of tools, these projects become harder and harder to maintain as data volumes increase and a desire for real-time pushes technology to its limit. New projects also struggle as new challenges of scale invalidate previous assumptions.
This talk will discuss some patterns that we see at Databricks that companies leverage to succeed with their data science projects. Key takeaways will be:
– Striving for simplicity
– Removing cognitive load for you and your team
– Working with data, big and small
– Effectively leveraging the ecosystem of tools to be successful
Study for big data analysis design modelJoon ho Park
. Recently Big data analysis is to analyze multiple data, and services that result. Big data services, advance the design of the analysis model, data collection, development, to the stage, such as the results derived. Data collection, development, result derivation, it is possible to use a common development methodology of the same standard and development business. However, the design of the analysis model has not yet been standardized. The design of most of the analysis model, the analysis of the CBD methodology, are using the design stage and agile methodology. In the present study, to present a strategy for the design of successful analysis model. For this reason, to analyze the business using the methodology of the CBD methodology and agile, to present a design proposal of the analysis model.
IT Governance & Management in Healthcare Organizations: Part 1 (October 19, 2...Nawanan Theera-Ampornpunt
Presented at the Master of Science and Doctor of Philosophy Programs in Data Science for Healthcare and Clinical Informatics, Department of Clinical Epidemiology and Biostatistics, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Bangkok, Thailand on October 19, 2020
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Find more Data-Ed webinars here: www.datablueprint.com
Data-Ed Online: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Takeaways:
Understanding how to contribute to organizational challenges beyond traditional data architecting
How to utilize data architectures in support of business strategy
Understanding foundational data architecture concepts based on the DAMA DMBOK
Data architecture guiding principles & best practices
IT Governance & Management in Healthcare Organizations: Part 1 (October 16, 2...Nawanan Theera-Ampornpunt
Presented at the Data Science for Healthcare Graduate Programs, Section for Clinical Epidemiology and Biostatistics, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Bangkok, Thailand on October 16, 2019
Information Technology Management in Healthcare Organizations: Part 1 (Octobe...Nawanan Theera-Ampornpunt
Presented at the Master of Science and Doctor of Philosophy Programs in Data Science for Healthcare and Clinical Informatics, Department of Clinical Epidemiology and Biostatistics, Faculty of Medicine Ramathibodi Hospital, Mahidol University, Bangkok, Thailand on October 20, 2021
Data Architecture is foundational to an information-based operational environment. Without proper structure and efficiency in organization, data assets cannot be utilized to their full potential, which in turn harms bottom-line business value. When designed well and used effectively, however, a strong Data Architecture can be referenced to inform, clarify, understand, and resolve aspects of a variety of business problems commonly encountered in organizations.
The goal of this webinar is not to instruct you in being an outright Data Architect, but rather to enable you to envision a number of uses for Data Architectures that will maximize your organization’s competitive advantage. With that being said, we will:
Discuss Data Architecture’s guiding principles and best practices
Demonstrate how to utilize Data Architecture to address a broad variety of organizational challenges and support your overall business strategy
Illustrate how best to understand foundational Data Architecture concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
High Performance Data Analytics and a Java Grande Run TimeGeoffrey Fox
There is perhaps a broad consensus as to important issues in practical parallel computing as applied to large scale simulations; this is reflected in supercomputer architectures, algorithms, libraries, languages, compilers and best practice for application development.
However the same is not so true for data intensive even though commercially clouds devote many more resources to data analytics than supercomputers devote to simulations.
Here we use a sample of over 50 big data applications to identify characteristics of data intensive applications and to deduce needed runtime and architectures.
We propose a big data version of the famous Berkeley dwarfs and NAS parallel benchmarks.
Our analysis builds on the Apache software stack that is well used in modern cloud computing.
We give some examples including clustering, deep-learning and multi-dimensional scaling.
One suggestion from this work is value of a high performance Java (Grande) runtime that supports simulations and big data
Similar to Data Mining of Project Management Data: An Analysis of Applied Research Studies_pptx (20)
Supplier and buyer driven channels in a two-stage supply chainGurdal Ertek
We explore the impact of power structure on price, sensitivity of market price, and profits in a two-stage supply chain with single product, supplier and buyer, and a price sensitive market. We develop and analyze the case where the supplier has dominant bargaining power and the case where the buyer has dominant bargaining power. We consider a pricing scheme for the buyer that involves both a multiplier and a markup. We show that it is optimal for the buyer to set the markup to zero and use only a
multiplier. We also show that the market price and its sensitivity are higher when operational costs (namely distribution and inventory) exist. We observe that the sensitivity of the market price increases non-linearly as the wholesale price increases, and derive a lower bound for it. Through experimental
analysis, we show that marginal impact of increasing shipment cost and carrying charge (interest rate) on prices and profits are decreasing in both cases. Finally, we show that there exist problem instances where the buyer may prefer supplier-driven case to markup-only buyer-driven and similarly problem
instances where the supplier may prefer markup-only buyer-driven case to supplier-driven.
http://research.sabanciuniv.edu.
Simulation modeling for quality and productivity in steel cord manufacturingGurdal Ertek
We describe the application of simulation modeling to estimate and improve quality and productivity performance of a steel cord manufacturing system. We describe the typical steel
cord manufacturing plant, emphasize its distinguishing characteristics, identify various production settings and discuss applicability of simulation as a management decision support tool. Besides presenting the general structure of the developed simulation model, we focus on wire fractures, which can be an important source of system disruption.
http://research.sabanciuniv.edu.
Visual and analytical mining of sales transaction data for production plannin...Gurdal Ertek
Recent developments in information technology paved the way for the collection of large amounts of data pertaining to various aspects of an enterprise. The greatest challenge faced in
processing these massive amounts of raw data gathered turns out to be the effective management of data with the ultimate purpose of deriving necessary and meaningful information
out of it. The following paper presents an attempt to illustrate the combination of visual and analytical data mining techniques for planning of marketing and production activities. The
primary phases of the proposed framework consist of filtering, clustering and comparison steps implemented using interactive pie charts, K-Means algorithm and parallel coordinate plots
respectively. A prototype decision support system is developed and a sample analysis session is conducted to demonstrate the applicability of the framework.
http://research.sabanciuniv.edu.
In crossdocking, the inbound materials coming in trucks to the
crossdock facility are directed to outbound doors and are directly loaded into trucks that will perform shipment, or are staged for a very brief time period before loading. Crossdocking has a great potential to bring savings in logistics: For example, most of the logistics success of Wal-Mart, the world’s leading retailer, is attributed to crossdocking.In this paper,the types of
crossdocking are identified, the situations and industries where crossdocking is applicable are explained, prerequisites, advantages and drawbacks are listed, and implementation issues are discussed. Finally a case study that describes the crossdocking applications of a 3rd party logistics firm is
presented.
http://research.sabanciuniv.edu.
Application of local search methods for solving a quadratic assignment proble...Gurdal Ertek
This paper discusses the design and application of local search methods to a real-life application at a steel cord manufacturing plant. The case study involves a layout problem that can be represented as a Quadratic Assignment Problem (QAP). Due to the nature of the manufacturing process, certain machinery need to be allocated in close proximity to each other. This issue is incorporated into the objective function through assigning high penalty costs to the unfavorable allocations. QAP belongs to one of the most difficult class of combinatorial optimization problems, and is not solvable to optimality as the number of facilities increases. We implement the well-known local
search methods, 2-opt, 3-opt and tabu search. We compare the solution performances of the methods to the results obtained from the NEOS server, which provides free access to many optimization solvers on the internet.
http://research.sabanciuniv.edu.
Financial benchmarking of transportation companies in the New York Stock Exch...Gurdal Ertek
In this paper, we present a benchmarking study of industrial transportation companies traded in the New York Stock Exchange (NYSE). There are two distinguishing aspects of our study: First, instead of using operational data for the input and the output items of the developed Data Envelopment Analysis (DEA) model, we use financial data of the companies that are readily available on the Internet. Secondly, we visualize the efficiency scores of the companies in relation to the subsectors and the number of employees. These visualizations enable us to discover interesting insights about the companies within each subsector, and about subsectors in comparison to each other. The visualization approach that we employ can be used in any DEA study that contains subgroups within a group. Thus, our paper also contains a methodological contribution.
http://research.sabanciuniv.edu.
Development of an Interactive Simulation of Steel Cord Manufacturing for Indu...Gurdal Ertek
We developed an interactive simulation program to be used in industrial engineering education, based on an earlier simulation study of a steel cord manufacturing plant. In the class project, the students are asked to design strategies/algorithms for finding the optimal values of operational decision variables by using the program.
http://research.sabanciuniv.edu.
Application of the analytic hierarchy process (AHP) for selection of forecast...Gurdal Ertek
In this paper, we described an application of the Analytic Hierarchy Process (AHP) for the ranking and selection of forecasting software. AHP is a multi-criteria decision making (MCDM) approach, which is based on the pair-wise comparison
of elements of a given set with respect to multiple criteria. Even though there are applications of the AHP to software selection problems, we have not encountered a study that involves forecasting software. We started our analysis by filtering
among forecasting software that were found on the Internet by undergraduate students as a part of a course project. Then we processed a second filtering step, where we reduced the number of software to be examined even further. Finally we
constructed the comparison matrices based upon the evaluations of three “semiexperts”, and obtained a ranking of forecasting software of the selected software using the Expert Choice software. We report our findings and our insights, together with the results of a sensitivity analysis.
http://research.sabanciuniv.edu.
Spreadsheet Optimization of Flow for WIP Goods at a Yarn and Tire Cord Manufa...Gurdal Ertek
We developed a spreadsheet optimization model to find the optimal material handling policy at a leading yarn and tire cord manufacturer’s facility. The focus of the study is to optimize the movement of looms of yarn and tire cord between two major manufacturing steps. The considered flow network within the
manufacturing facility is a miniature two-stage supply chain, including the middle layer of depots. Movement of material takes place through forklift trucks and cranes, which impose electricity, labor, maintenance, and depreciation costs. After
introducing the company, the facility, and the project goals, we present the underlying linear programming (LP) model and our spreadsheet implementation. The results of our model suggest an annual saving of approximately $22,000 for the facility.
http://research.sabanciuniv.edu.
Optimizing waste collection in an organized industrial region: A case studyGurdal Ertek
In this paper we present a case study which involves the design of a supply chain network for industrial waste collection. The problem is to transport metal waste from 17 factories to containers and from containers to a disposal center (DC) at an
organized region of automobile parts suppliers. We applied the classic mixedinteger programming (MIP) model for the two-stage supply chain to the solution of this problem. The visualization of the optimal solution provided us with several
interesting insights that would not be easily discovered otherwise.
http://research.sabanciuniv.edu.
Demonstrating warehousing concepts through interactive animationsGurdal Ertek
In this paper, we report development of interactive computer animations to demonstrate warehousing concepts, providing a virtual environment for learning. Almost every company, regardless of its industry, holds inventory of goods in its
warehouse(s) to respond to customer demand promptly, to coordinate supply and demand, to realize economies of scale in manufacturing or processing, to add value to its products and to reduce response time. Design, analysis, and improvement of warehouse operations can yield significant savings for a company. Warehousing science can be considered as an important field within the industrial engineering discipline. However, there is very little educational material (including web based media), and only a handful of books available in this field.
We believe that the animations that we developed will significantly contribute to the understanding of warehousing concepts, and enable tomorrow’s practitioners to grasp the fundamentals of managing warehouses.
http://research.sabanciuniv.edu.
Decision Support For Packing In WarehousesGurdal Ertek
Packing problems deal with loading of a set of items (ob-
jects) into a set of boxes (containers) in order to optimize a performance criterion under various constraints. With the advance of RFID technologies and investments in IT infrastructures companies now have Access to the necessary data that can be utilized in cost reduction of packing processes. Therefore bin packing and container loading problems are be-
coming more popular in recent years. In this research we propose a beam search algorithm to solve a packing problem that we encountered in a real world project. The 3D-MBSBPP (Multiple Bin Sized Bin Packing Problem) that we present and solve has not been analyzed in literatüre before, to the best of our knowledge. We present the performance of our proposed beam search algorithm in terms of both cost and computational
time in comparison to a greedy algorithm and a tree search enumeration algorithm.
http://research.sabanciuniv.edu.
A framework for visualizing association mining resultsGurdal Ertek
Association mining is one of the most used data mining tech-
niques due to interpretable and actionable results. In this study we propose a framework to visualize the association mining results, specifically frequent itemsets and association rules, as graphs. We demonstrate the applicability and usefulness of our approach through a Market Basket Analysis (MBA) case study where we visually explore the data mining results for a supermarket data set. In this case study we derive several
interesting insights regarding the relationships among the items and suggest how they can be used as basis for decision making in retailing.
http://research.sabanciuniv.edu.
Application of the cutting stock problem to a construction company: A case studyGurdal Ertek
This paper presents an application of the well-known cutting stock problem to a construction firm. The goal of the 1Dimensional (1D) cutting stock problem is to cut the bars of desired lengths in required quantities from longer bars of given
length. The company for which we carried out this study encounters 1D cutting stock problem in cutting steel bars (reinforcement bars) for its construction projects. We have developed several solution approaches to solving the company’s problem: Building and solving an integer programming (IP) model in a modeling environment, developing our own software that uses a mixed integer programming (MIP) software library, and testing some of the commercial software packages available on the internet. In this paper, we summarize our experiences with all the three approaches. We also present a benchmark of existing commercial software packages, and some critical insights. Finally, we suggest a visual approach for increasing performance in solving the cutting stock problem and demonstrate the applicability of this approach using the company’s data on two construction projects.
http://research.sabanciuniv.edu.
Benchmarking the Turkish apparel retail industry through data envelopment ana...Gurdal Ertek
This paper presents a benchmarking study of the Turkish apparel retailing industry. We have applied the Data Envelopment Analysis (DEA) methodology to determine the efficiencies of the companies in the industry. In the DEA model the number of stores, number of corners, total sales area and number of employees were included as inputs and annual sales revenue was included as the output. The efficiency scores obtained through DEA were visualized for gaining insights about
the industry and revealing guidelines that can aid in strategic decision making.
http://research.sabanciuniv.edu.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Data Mining of Project Management Data: An Analysis of Applied Research Studies_pptx
1. Data Mining of Project Management Data:
An Analysis of Applied Research Studies
1
Gürdal Ertek Allan N. Zhang Sobhan Asian Murat M Tunc Omer Tanrikulu
3. Project
• “A temporary endeavor
• (with a definite beginning and definite end),
• with progressive elaboration (developing in
steps, continuing in increments),
• undertaken to create a unique
• product,
• service, or
• result”.
3
4. Project Management (PM)
• “Application of
• knowledge,
• skills,
• tools, and
• techniques
• to project activities
• to meet project requirements” [1].
• Very important, because projects are
• undertaken at all levels of the organization,
• in almost any industry, and
• can have long-term effects.
4
5. Project Management (PM)
• Vast literature on project management
• Specialized academic journals,
• Professional institutions,
• Project Management Institute (PMI),
• dedicated to project management field.
5
6. Project Management (PM)
• However, real world projects usually
• fall behind the performance goals or
• FAIL frequently.
6Source: http://calleam.com/WTPF/
7. Project Management (PM)
• McKinsey & Company study of 5,400 large scale
information technology (IT) projects:
• Large IT projects run
• 45% over budget and
• 7% over time, while
• delivering 56% less value than predicted.
• Even more critically,
• 17% of large IT projects FAIL so big to threaten the
existence of the company.
7
8. Data in Projects (1 of 2)
• Project managers and planners make use of data,
while new data is generated as a project
progresses.
• Real world projects are becoming increasingly
complex, and are involving larger amount of
data.
• Big strategic projects, such as the production
of a satellite, already generate amounts of data
that can be classified as big data.
8
9. Data in Projects (2 of 2)
• Data
• in databases
• post-project reviews
• can be
• major source of actionable insights and competitive
advantage.
• Multitude of studies
• data mining (DM) techniques for
• analyzing data coming from project management
(PM).
• "DM for PM"
9
10. Data Mining
• Data mining (DM)
• growing field of computer science
• discovery of actionable insights from –typically large
and complex- data
• tap into the information hidden in databases and
unstructured documents
• use data for advantage.
10
11. Research Gap | Topic
• No research on the survey of literature on
• data mining applications for project management.
• While use of data mining in manufacturing was
• surveyed in [9].
• Our Research:
• investigation of the literature on "DM for PM",
• results and analysis.
• gaps in the current literature
• opportunities for future research.
• Goal:
• understanding of data mining (DM) for project
management (PM) researchers and project managers,
• as the world is moving towards the new age of big data. 11
17. Research Methods (1 of 2)
• Data Mining
• growing field of computer science and informatics
• aims at discovering new and useful information and
knowledge from data.
• multitude of analytical methods (and algorithms)
• each method or combination of methods are most
suitable for a given data with unique characteristics.
17
18. Research Methods (2 of 2)
• Association Mining
• data mining method for
• identifying associations between
• elements (items) of a set (set of items),
• based on how these elements appear in
• multiple subsets (transactions) of the set.
• gives as output
• list of itemsets appearing together frequently in
transactions (frequent itemsets), and
• rules that describe how these associations affect each other
(association rules).
• An association rule is a rule in the form
“IF [Antecedent A] THEN [Consequent B]”
(or simply as “A⇒B”).
18
25. Conclusions (1 of 15)
25
• 41 of the 116 reviewed papers are coming from the
construction industry,
• showing the significance of construction industry from
• not only from a project management (PM) perspective,
• but also from data mining (DM) and information technology (IT)
perspectives.
26. Conclusions (2 of 15)
26
• Other frequently encountered industries in the papers are
• information and communication
• manufacturing.
27. Conclusions (3 of 15)
27
• Papers using data from
• United States (19 papers) are most frequent, followed by those
that use data from
• Taiwan (14 papers) and
• China (5 papers).
28. Conclusions (4 of 15)
28
• Most frequent objectives are
• cost minimization,
• cost estimation,
• makespan and
• time minimization.
29. Conclusions (5 of 15)
29
• Visualization is the most popular data mining method, and is
• followed by statistical analysis.
• The application of association rule mining and text mining
seems least popular,
• illustrating the opportunity to conduct research that uses
these methods and/or develops new algorithms within these
methods, especially for manufacturing.
30. Conclusions (6 of 15)
30
• The most popular software tool is the
• SPSS statistics/data mining software, followed by
• MATLAB and
• WEKA.
31. Conclusions (7 of 15)
31
• An overwhelming percentage (88.8%) of the papers used
data from the real world, which is very favorable.
32. Conclusions (8 of 15)
32
• 85.3% of the papers used only existing methods,
• rather than developing new data mining methods for the project
management domain,
• or being applied in the project management domain.
• This shows an important opportunity for future research for
• developing new data mining methods
• for the project management domain.
33. Conclusions (9 of 15)
33
• 82.9% of the papers did not present the development of a
decision support system (DSS), which suggests that
• future research can involve development of DSS.
34. Conclusions (10 of 15)
34
• 78.4% of the papers looked into single project data,
• showing a gap, as well as opportunity to
• conduct research on multi-project management.
35. Conclusions (11 of 15)
35
• The data type in the papers was mainly (78%) single project
data, suggesting gap and opportunity to conduct
• more research with multiple-project data.
36. Conclusions (12 of 15)
36
• Research on multi-project data where projects share
resources is very scarce (9%), suggesting that
• research on multi-project data can especially focus on the case
where resources are shared.
Money Manpower Equipment
Facilities Materials Information/technology
37. Conclusions (13 of 15)
37
• More research can be done for
• operational-level projects and
• strategic-level projects,
• due to the gap and opportunity on projects at these levels.
38. Conclusions (14 of 15)
38
• There is opportunity to do more research that involves
• manufacturing, as well as
• public,
• defense, and
• scientific projects,
and projects in
• health,
• insurance, and
• energy industries.
39. Conclusions (15 of 15)
39
• Papers where decision support systems (DSS) were
developed are
• four times more likely to also contain the
• development of a new method.
• So any research where DSS or a new method is developed is
more likely to contain (and expected to contain by the
reviewers) the other.
42. Acknowledgement
• Data Cleaning & Analysis
• Şevki Murat Ayan
• Onur Aksöyek
• Ece Kurtaraner
• Mete Sevinç
• Byung-Geun Cho
• Research Grant
• Abu Dhabi University
42