meta360 - enterprise data governance and metadata managementBojana Ciric
meta360 is an enterprise scale, industry agnostic, the state-of-the-art data governance and metadata management tool which provides an easy way to collect and manage all relevant business and technical metadata from your enterprise data environment, as well as powerful visualization capabilities to easily navigate through metadata content and use the information on the most effective way. meta360 is industry agnostic and can be used as a key component in various data management initiatives, including, but not limited to: data governance,data lineage, metadata management, data quality, MDM, data integration, analytics, etc.
Features:
1) Innovative, matured and proven approach for data governance operationalization
2)Industry agnostic, can be used in various industries (FSI, communications, life science, etc.)
3)Easy to implement – up and running within 6 weeks, even for the large organizations
4)Cloud based (Amazon Cloud) – significantly reduces operational costs
5)Easy content contribution – CSV and JSON file import, manual entry (can be used as primary tool for particular concept types)
6)Exceptional user experience – visually attractive and easy-to-use for both, business and technical users.
6)Responsive, works on all devices
7)meta 360 is built by using MEAN stack
World of Watson 2016 - Implementing data scienceKeith Redman
Wikipedia defines Data Science as an interdisciplinary field about processes and systems to extract knowledge or insights from data in various forms, either structured or unstructured, which is a continuation of some of the data analysis fields such as statistics, machine learning, data mining, and predictive analytics… It then goes on to say Data Science is not to be confused with Information Science which is not to be confused with Library Science which is not to be confused with….. I think you get the point.
Are you confused about the benefits of Data Science, or have you embraced Data Science and are now struggling to enable your Scientists? Check out the sessions on Data Science and especially take a look at the new Data Science Experience.
QU Summer school 2020 speaker Series - Session 7
A conversation with Quants, Thinkers and Innovators all challenged to innovate in turbulent times!
Join QuantUniversity for a complimentary summer speaker series where you will hear from Quants, innovators, startups and Fintech experts on various topics in Quant Investing, Machine Learning, Optimization, Fintech, AI etc.
Managing Machine Learning Models in the Financial Industry
Lecture 1: Model Risk Management for AI and Machine Learning
Artificial intelligence and machine learning are part of today’s modeler’s toolbox for building challenger models and new innovative models that address business needs. However, AI presents new and unique challenges for risk management, particularly for assessing, controlling, and managing model risk for models of limited transparency. Another key consideration is the speed at which these models can be developed, validated, and then deployed into productive use to be competitive adhering to a robust model risk management program. This talk will highlight best practices for integrating AI into model risk practices and showcase examples across the model lifecycle.
The Softer Skills analysts need to succeed in their careersPaul Laughlin
A talk that I gave to the UK's Operational Research Society's annual Analytics & AI Summit 2021 (#AS21). Focusing on the Contracting & Delivering parts of this training model.
Presentation to Analytics Network of the OR Society Nov 2020Paul Laughlin
Presentation on 'The Softer Skills that Analysts need' presented by Paul Laughlin at a virtual event run for the Analytics Network group within the UK OR Society. Exploring Paul's 9 Step Model for effective analysis & explaining how Softer Skills are essential throughout that workflow.
The People Skills analysts need to succeed in their careersPaul Laughlin
Slides from a webinar that I presented. Sharing some of the highlights of my one-day online training course on People Skills for Analysts, including my 9-step-model for effective analysis. Further details available from: http://laughlinconsultancy.com
The Softer Skills that analysts need (beyond Data Visualisation)Paul Laughlin
A talk I gave at #DataVizLive online event in Nov 2020. Introducing the Laughlin Consultancy 9-step model for Softer Skills needed by Analysts & previewing some of those steps (beyond data visualisation & storytelling skills).
meta360 - enterprise data governance and metadata managementBojana Ciric
meta360 is an enterprise scale, industry agnostic, the state-of-the-art data governance and metadata management tool which provides an easy way to collect and manage all relevant business and technical metadata from your enterprise data environment, as well as powerful visualization capabilities to easily navigate through metadata content and use the information on the most effective way. meta360 is industry agnostic and can be used as a key component in various data management initiatives, including, but not limited to: data governance,data lineage, metadata management, data quality, MDM, data integration, analytics, etc.
Features:
1) Innovative, matured and proven approach for data governance operationalization
2)Industry agnostic, can be used in various industries (FSI, communications, life science, etc.)
3)Easy to implement – up and running within 6 weeks, even for the large organizations
4)Cloud based (Amazon Cloud) – significantly reduces operational costs
5)Easy content contribution – CSV and JSON file import, manual entry (can be used as primary tool for particular concept types)
6)Exceptional user experience – visually attractive and easy-to-use for both, business and technical users.
6)Responsive, works on all devices
7)meta 360 is built by using MEAN stack
World of Watson 2016 - Implementing data scienceKeith Redman
Wikipedia defines Data Science as an interdisciplinary field about processes and systems to extract knowledge or insights from data in various forms, either structured or unstructured, which is a continuation of some of the data analysis fields such as statistics, machine learning, data mining, and predictive analytics… It then goes on to say Data Science is not to be confused with Information Science which is not to be confused with Library Science which is not to be confused with….. I think you get the point.
Are you confused about the benefits of Data Science, or have you embraced Data Science and are now struggling to enable your Scientists? Check out the sessions on Data Science and especially take a look at the new Data Science Experience.
QU Summer school 2020 speaker Series - Session 7
A conversation with Quants, Thinkers and Innovators all challenged to innovate in turbulent times!
Join QuantUniversity for a complimentary summer speaker series where you will hear from Quants, innovators, startups and Fintech experts on various topics in Quant Investing, Machine Learning, Optimization, Fintech, AI etc.
Managing Machine Learning Models in the Financial Industry
Lecture 1: Model Risk Management for AI and Machine Learning
Artificial intelligence and machine learning are part of today’s modeler’s toolbox for building challenger models and new innovative models that address business needs. However, AI presents new and unique challenges for risk management, particularly for assessing, controlling, and managing model risk for models of limited transparency. Another key consideration is the speed at which these models can be developed, validated, and then deployed into productive use to be competitive adhering to a robust model risk management program. This talk will highlight best practices for integrating AI into model risk practices and showcase examples across the model lifecycle.
The Softer Skills analysts need to succeed in their careersPaul Laughlin
A talk that I gave to the UK's Operational Research Society's annual Analytics & AI Summit 2021 (#AS21). Focusing on the Contracting & Delivering parts of this training model.
Presentation to Analytics Network of the OR Society Nov 2020Paul Laughlin
Presentation on 'The Softer Skills that Analysts need' presented by Paul Laughlin at a virtual event run for the Analytics Network group within the UK OR Society. Exploring Paul's 9 Step Model for effective analysis & explaining how Softer Skills are essential throughout that workflow.
The People Skills analysts need to succeed in their careersPaul Laughlin
Slides from a webinar that I presented. Sharing some of the highlights of my one-day online training course on People Skills for Analysts, including my 9-step-model for effective analysis. Further details available from: http://laughlinconsultancy.com
The Softer Skills that analysts need (beyond Data Visualisation)Paul Laughlin
A talk I gave at #DataVizLive online event in Nov 2020. Introducing the Laughlin Consultancy 9-step model for Softer Skills needed by Analysts & previewing some of those steps (beyond data visualisation & storytelling skills).
[Notes] Customer 360 Analytics with LEO CDPTrieu Nguyen
Part 1: Why should every business need to deploy a CDP ?
1. Big data is the reality of business today
2. What are technologies to manage customer data ?
3. The rise of first-party data and new technologies for Digital Marketing
4. How to apply USPA mindset to build your CDP for data-driven business
Part 2: How to use LEO CDP for your business
1. Core functions of LEO CDP for marketers and IT managers
2. Data Unification for Customer 360 Analytics
3. Data Segmentation
4. Customer Personalization
5. Customer Data Activation
Part 3: Case study in O2O Retail and Ecommerce
1. How to build customer journey map for ecommerce and retail
2. How to do customer analytics to find ideal customer profiles
The ideal customer profile in a B2B context
The ideal customer profile in a B2C context
3. Manage product catalog for customer personalization
4. Monitoring Data of Customer Experience (CX Analytics)
CX Data Flow
CX Rating plugin is embedded in the website, to collect feedback data
An overview of CX Report
A CX Report in a customer profile
5. Monitoring data with real-time event tracking reports
Event Data Flow
Summary Event Data Report
Event Data Report in a Customer Profile
Part 4: How to setup an instance of LEO CDP for free
1. Technical architecture
2. Server infrastructure
3. Setup middlewares: Nginx, ArangoDB, Redis, Java and Python
Network requirements
Software requirements for new server
ArangoDB
Nginx Proxy
SSL for Nginx Server
Java 8 JVM
Redis
Install Notes for Linux Server
Clone binary code for new server
Set DNS hosts for LEO CDP workers
4. Setup data for testing and system verification
Part 5: Summary all key ideas
How to use the EU e-Competence Framework together with the Body of Knowledge Edison when implementing a new workforce model. In this use case Workforce Transformation, Curricula, Career paths and assessments. Human Resource Development tooling.
Advanced data visualization (ADV) is a rapidly emerging concept in business and society and has a lofty goal of transforming data into information. But how do we get there?
Analytics & Data Strategy 101 by Deko DimeskiDeko Dimeski
- Understand why each company needs solid analytics and data strategy & capabilities
- Typical data problems each company experiences, regardless of the scale
- Core competences and roles
- Analytics products and artefacts
- Analytics Usecases
LoQutus helps organisations to innovate with analytics and to get insights with data visualisation. We also build large scale data layers to enable interaction with core data, and develop data-driven applications to deliver the insights our customers need. During this session we’ll share what we have learned along the way. We’ll show you our framework for self-service analytics & insights, and some successful case studies.
Business analytics and its basic concepts
The presentation can help you to understand the basic concepts of business analytics, process of analytics, scope and nature of analytics, types of analytics and advantages of analytics.
Best Data Science Hybrid Course in Pune
Data Science, in its simpler terms, is about generating critical business value from the data through various creative ways. It can also be defined as a mix of data research, algorithms, and technology to solve complex analytical issues. Data is being generated by Companies at an exponential pace. The usable Data form can be different for different sections of people working in an organization.
Data Science Classes help us to explore the data to a granular form and find the needed insights. Data Science is about being analytical or inquisitive wherein asking new questions, doing further explorations, and continuing learning is a part of the job for Data Scientists.
According to Harward Business Review, Data Scientist is the Sexiest Job of the 21st Century.
According to Forbes, IBM Predicts Demand For Data Scientists Will Soar 28% By 2020
GET FRONTLINE DATA SCIENCE TRAINING IN PUNE AT 3RI TECHNOLOGIES
Data Science is a trending niche, for it promises notable mileage for the business economy! It is rather ironic that data which was considered a burden to manage and store only about a few decades ago is now viewed as a resource; courtesy of course to data scientists. They have brought about a paradigmatic change through their skills which allow them to derive the value from raw data. It is important to mention that ‘Raw Data’ is clueless to most laymen, including the high echelons in business management; but when processed through Data Science Tools, it renders value that is precious and immense for the decision-makers and salesmen. They are all riding on the Professionalism of the Data Scientists and this generates the demand of the latter! 3RI Technologies is the leading institution offering Data Science Classes in Pune and fresh graduates as well as Working Professionals can enroll for it.
WHAT IS DATA SCIENCE?
Today, Data Science is a much-talked subject and its significance is being deliberated among the business managers who are eager to hire a brilliant professional onboard their firm. Data Science is a milieu space that is shared by the distinct yet related domains of statistics & applicative mathematics, computer programming frameworks and tools, data metrics, and analytics. Machine Learning & associated automation underpins all the above-listed fields, almost as a generic derivative; because it is through this channel that the good results are accrued in favor of the business clients. What are these good results? Let’s talk about them!
Trending smart services that are propelling businesses around the world such as SEO, SMO, SMM, SEM and CRM, all revolve around the ability to generate leads of authentic value for the commerce banners. The web developers have been doing well through their professional conduct for their clients but they in turn actively seek the ‘Meaningful Data’ about the existing and potential customers, the market trends, and the competition figures of the biz rivals. Here, Data Sc
40 ° advises and supports companies and institutions to generate real added value from data and to generate data-driven innovations and new business models. We help to reinvent your business with data. 40 ° is the expert for data driven business transformation
An efficient data science team is crucial for deriving value from the humongous data a business collect. Learn how the data science team can help in this regard.
How Data Virtualization Puts Enterprise Machine Learning Programs into Produc...Denodo
Watch full webinar here: https://bit.ly/3offv7G
Presented at AI Live APAC
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spend most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Watch this on-demand session to learn how companies can use data virtualization to:
- Create a logical architecture to make all enterprise data available for advanced analytics exercise
- Accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- Integrate popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc.
Advanced Analytics and Machine Learning with Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3aXysas
Advanced data science techniques, like machine learning, have proven to be extremely useful to derive valuable insights from your data. Data Science platforms have become more approachable and user friendly. With all the advancements in the technology space, the Data Scientist is still spending most of the time massaging and manipulating the data into a usable data asset. How can we empower the data scientist? How can we make data more accessible, and foster a data sharing culture?
Join us, and we will show you how Data Virtualization can do just that, with an agile and AI/ML laced data management platform. It can empower your organization, foster a data sharing culture, and simplify the life of the data scientist.
Watch this webinar to learn:
- How data virtualization simplifies the life of the data scientist, by overcoming data access and manipulation hurdles.
- How integrated Denodo Data Science notebook provides for a unified environment
- How Denodo uses AI/ML internally to drive the value of the data and expose insights
- How customers have used Data Virtualization in their Data Science initiatives.
Big Data and Analytics in your Organisation talk.pdfPaul Laughlin
My presentation to the ILA Digital Community Showcase event at the Vale Hotel in Glamorgan on 7 Oct 2022. Introducing the topic of Digital Transformation to healthcare leaders in Wales.
Data Leadership talk for CIIA March 2022.pdfPaul Laughlin
Slides presented to the Chartered Institute of Internal Auditors on the challenge of being a data leader & the skills needed for such leaders (and their teams) to succeed.
You can read more about this event here:
[Notes] Customer 360 Analytics with LEO CDPTrieu Nguyen
Part 1: Why should every business need to deploy a CDP ?
1. Big data is the reality of business today
2. What are technologies to manage customer data ?
3. The rise of first-party data and new technologies for Digital Marketing
4. How to apply USPA mindset to build your CDP for data-driven business
Part 2: How to use LEO CDP for your business
1. Core functions of LEO CDP for marketers and IT managers
2. Data Unification for Customer 360 Analytics
3. Data Segmentation
4. Customer Personalization
5. Customer Data Activation
Part 3: Case study in O2O Retail and Ecommerce
1. How to build customer journey map for ecommerce and retail
2. How to do customer analytics to find ideal customer profiles
The ideal customer profile in a B2B context
The ideal customer profile in a B2C context
3. Manage product catalog for customer personalization
4. Monitoring Data of Customer Experience (CX Analytics)
CX Data Flow
CX Rating plugin is embedded in the website, to collect feedback data
An overview of CX Report
A CX Report in a customer profile
5. Monitoring data with real-time event tracking reports
Event Data Flow
Summary Event Data Report
Event Data Report in a Customer Profile
Part 4: How to setup an instance of LEO CDP for free
1. Technical architecture
2. Server infrastructure
3. Setup middlewares: Nginx, ArangoDB, Redis, Java and Python
Network requirements
Software requirements for new server
ArangoDB
Nginx Proxy
SSL for Nginx Server
Java 8 JVM
Redis
Install Notes for Linux Server
Clone binary code for new server
Set DNS hosts for LEO CDP workers
4. Setup data for testing and system verification
Part 5: Summary all key ideas
How to use the EU e-Competence Framework together with the Body of Knowledge Edison when implementing a new workforce model. In this use case Workforce Transformation, Curricula, Career paths and assessments. Human Resource Development tooling.
Advanced data visualization (ADV) is a rapidly emerging concept in business and society and has a lofty goal of transforming data into information. But how do we get there?
Analytics & Data Strategy 101 by Deko DimeskiDeko Dimeski
- Understand why each company needs solid analytics and data strategy & capabilities
- Typical data problems each company experiences, regardless of the scale
- Core competences and roles
- Analytics products and artefacts
- Analytics Usecases
LoQutus helps organisations to innovate with analytics and to get insights with data visualisation. We also build large scale data layers to enable interaction with core data, and develop data-driven applications to deliver the insights our customers need. During this session we’ll share what we have learned along the way. We’ll show you our framework for self-service analytics & insights, and some successful case studies.
Business analytics and its basic concepts
The presentation can help you to understand the basic concepts of business analytics, process of analytics, scope and nature of analytics, types of analytics and advantages of analytics.
Best Data Science Hybrid Course in Pune
Data Science, in its simpler terms, is about generating critical business value from the data through various creative ways. It can also be defined as a mix of data research, algorithms, and technology to solve complex analytical issues. Data is being generated by Companies at an exponential pace. The usable Data form can be different for different sections of people working in an organization.
Data Science Classes help us to explore the data to a granular form and find the needed insights. Data Science is about being analytical or inquisitive wherein asking new questions, doing further explorations, and continuing learning is a part of the job for Data Scientists.
According to Harward Business Review, Data Scientist is the Sexiest Job of the 21st Century.
According to Forbes, IBM Predicts Demand For Data Scientists Will Soar 28% By 2020
GET FRONTLINE DATA SCIENCE TRAINING IN PUNE AT 3RI TECHNOLOGIES
Data Science is a trending niche, for it promises notable mileage for the business economy! It is rather ironic that data which was considered a burden to manage and store only about a few decades ago is now viewed as a resource; courtesy of course to data scientists. They have brought about a paradigmatic change through their skills which allow them to derive the value from raw data. It is important to mention that ‘Raw Data’ is clueless to most laymen, including the high echelons in business management; but when processed through Data Science Tools, it renders value that is precious and immense for the decision-makers and salesmen. They are all riding on the Professionalism of the Data Scientists and this generates the demand of the latter! 3RI Technologies is the leading institution offering Data Science Classes in Pune and fresh graduates as well as Working Professionals can enroll for it.
WHAT IS DATA SCIENCE?
Today, Data Science is a much-talked subject and its significance is being deliberated among the business managers who are eager to hire a brilliant professional onboard their firm. Data Science is a milieu space that is shared by the distinct yet related domains of statistics & applicative mathematics, computer programming frameworks and tools, data metrics, and analytics. Machine Learning & associated automation underpins all the above-listed fields, almost as a generic derivative; because it is through this channel that the good results are accrued in favor of the business clients. What are these good results? Let’s talk about them!
Trending smart services that are propelling businesses around the world such as SEO, SMO, SMM, SEM and CRM, all revolve around the ability to generate leads of authentic value for the commerce banners. The web developers have been doing well through their professional conduct for their clients but they in turn actively seek the ‘Meaningful Data’ about the existing and potential customers, the market trends, and the competition figures of the biz rivals. Here, Data Sc
40 ° advises and supports companies and institutions to generate real added value from data and to generate data-driven innovations and new business models. We help to reinvent your business with data. 40 ° is the expert for data driven business transformation
An efficient data science team is crucial for deriving value from the humongous data a business collect. Learn how the data science team can help in this regard.
How Data Virtualization Puts Enterprise Machine Learning Programs into Produc...Denodo
Watch full webinar here: https://bit.ly/3offv7G
Presented at AI Live APAC
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spend most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Watch this on-demand session to learn how companies can use data virtualization to:
- Create a logical architecture to make all enterprise data available for advanced analytics exercise
- Accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- Integrate popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc.
Advanced Analytics and Machine Learning with Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3aXysas
Advanced data science techniques, like machine learning, have proven to be extremely useful to derive valuable insights from your data. Data Science platforms have become more approachable and user friendly. With all the advancements in the technology space, the Data Scientist is still spending most of the time massaging and manipulating the data into a usable data asset. How can we empower the data scientist? How can we make data more accessible, and foster a data sharing culture?
Join us, and we will show you how Data Virtualization can do just that, with an agile and AI/ML laced data management platform. It can empower your organization, foster a data sharing culture, and simplify the life of the data scientist.
Watch this webinar to learn:
- How data virtualization simplifies the life of the data scientist, by overcoming data access and manipulation hurdles.
- How integrated Denodo Data Science notebook provides for a unified environment
- How Denodo uses AI/ML internally to drive the value of the data and expose insights
- How customers have used Data Virtualization in their Data Science initiatives.
Big Data and Analytics in your Organisation talk.pdfPaul Laughlin
My presentation to the ILA Digital Community Showcase event at the Vale Hotel in Glamorgan on 7 Oct 2022. Introducing the topic of Digital Transformation to healthcare leaders in Wales.
Data Leadership talk for CIIA March 2022.pdfPaul Laughlin
Slides presented to the Chartered Institute of Internal Auditors on the challenge of being a data leader & the skills needed for such leaders (and their teams) to succeed.
You can read more about this event here:
Do data leaders face unique challenges as leaders?Paul Laughlin
A leadership development talk given to private event for senior data leaders - it sparked some really useful conversations. To find out more please checkout: http://laughlinconsultancy.com
It's so important at this time to use trusted sources for your analysis & Data Viz, even in simple dashboards. This deck helps reveal some of the sources you can use.
The Softer Skills Analysts need to make an impactPaul Laughlin
25 min presentation given at London Business School, to the OR Society's Analytics Network. Summarising Laughlin Consultancy's 9 step model of Softer Skills for Analysts.
Sharing results of researching 1k UK consumers on their digital customer experience, to show need for positive response to GDPR & the key role of Identity Relationship Management.
How life-event data can improve & protect your marketing in a post-GDPR worldPaul Laughlin
Understanding the role of life-event data and life-event marketing in both improving marketing effectiveness and protecting consumer interest/reach following GDPR enforcement. Sponsored by Royal Mail Data Services and MyCustomer.com
Softer Skills workshop for #DTS16 event in EdinburghPaul Laughlin
Slides presented as keynote workshop at Data Talent Scotland 2016 event, brining together Data Science students, academics, employers & data industry for conversation & talent pipeline.
Presentation at Big Data & Analytics for Insurance 2016Paul Laughlin
Presentation on the Softer Skills that Data professionals also need to make an impact. Summary of key lessons from the very popular Consultancy Skills for Analysts training course from LaughlinConsultancy.com
Behavioural economics for The Financial Services Forum members conference 2015Paul Laughlin
A brief introduction to Behavioural Economics and suggestions for getting started with applying such hypotheses to improving customer communications. Presented at The Financial Services Forum annual members conference in March 2015.
An introduction to Laughlin Consultancy. Who we are, what we do & why companies need help with Customer Insight Leadership. Our Services include Consultancy, Training & Coaching. Happy to chat further if you want to maximise the value of your Customer Insight.
Presentation on Measuring Marketing Effectiveness which I presented at #CAI2015 in London on 25 Feb 2015. A review of that event can be found at: http://customerinsightleader.com/events/event-cai2015/
Gain Deeper Insights: Using analytics and research to generate insightsPaul Laughlin
An introduction to Customer Insight, explaining what it is & how to pull multiple technical skills together to generate customer insights which provoke action. Paul Laughlin shares learning from over a decade of creating and leading customer insight teams to make £10m difference p.a. to bottom lines.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
1. Paul Laughlin
Chief Blogger, CustomerInsightLeader.com
Founder, LaughlinConsultancy.com
“Everyday Data Science”
Masterclass
2. My background in creating data,
analytics & Data Science teams
2
❖ Created and lead data, analytics & data
science teams for all the major insurance
brands, products & channels used by
Lloyds Banking Group over 13 years
❖ Added over £11m incremental profit to
bottom line annually
❖ Pioneered work with FCA on Behavioural
Economics in comms
❖ Developed capability in team of 44 &
mentored next generation of leaders
3. The way spend my time now
reflects needs of Data Science
3
“Helping businesses make money from customer insight”
A selection of the clients I work with to embed these skills…
4. The 2nd biggest predictor of
whether you get value from this…
4
Goals:
What do you want to get
out of this morning?
5. 01 All around us, but let’s get clearer
What is it?
02 Some exciting & concerning applications
How is it being used?
04 How could you learn the coding needed?
Coding Opportunities
03 How can you spot new data & uses?
Data Opportunities
05 What is needed to get started?
Getting ready
5
09:45
10:30
11:30
12:00
12:30
11. But what is Data Science & where
has it come from?
11Source: CapGemini.com
12. Probably the most popular
definition of Data Science
12
Source: DrewConway.com Source: oralytics.com Source: datacommunitydc.org
13. Breadth of Skills needed is
recognised for Data Scientists
Drew Conway “The Data Science Venn Diagram” March 2013
<http://drewconway.com/ zia/2013/3/26/the-data-science-venn-diagram>
Page 11 of 59
Figure 3. ICTprocessstagesalignedwiththe organisationalproductionworkflow (as usedine-CF3.0)
Figure 4 illustrates the multi-purpose use of the European e-Competence Framework within ICT organisations.
The e-CF has a multidimensional structure and is flexible in using for different purposes, it can be easy adopted
for organisation specific model and roles. The e-CF3.0 is used for job-profiles definition in CWA 16458 (see [9]
and EDSF DSPP document [4]) that are linked to the organisational processes what creates limitations for cross-
organisational professional profiles and roles such as Data Scientist. However, combining competences from
different competence areas and using them as building blocks can allow flexible job-profiles definition. This
enables the derived job-profiles to be easily updated by changing set of competences related to profiles without
the need to restructure the entire profile.
Figure 4. e-CF3.0 structure and use for definition of the job profile definition and training needs.
EU EDISON Data Science Framework (2017)
13
14. but too much of focus is on
developing Technical Skills
EDSF Release 2: Part 1. Data Science Competence Framework (CF-DS)
Table 4.2. Identified Data Science skills related to the main Data Science competence groups
SDSDA
Data Science
Analytics
SDSENG
Data Science
Engineering
SDSDM
Data Management
SDSRM
Research Methods
and Project
Management
SDSBA
Business Analytics
SDSDA01
Use Machine Learning
technology,
algorithms, tools
(including supervised,
unsupervised, or
reinforced learning)
SDSENG01
Use systems and
software engineering
principles to
organisations
information system
design and development,
including requirements
design
SDSDM01
Specify, develop and
implement enterprise
data management and
data governance
strategy and
architecture, including
Data Management Plan
(DMP)
SDSRM01
Use research methods
principles in developing
data driven applications
and implementing the
whole cycle of data
handling
SDSBA01
and Business
Intelligence (BI)
methods for data
analysis; apply
cognitive
technologies and
relevant services
SDSDA02
Use Data Mining
techniques
SDSENG02
Use Cloud Computing
technologies and cloud
powered services design
for data infrastructure
and data handling
services
SDSDM02
Data storage systems,
data archive services,
digital libraries, and their
operational models
SDSRM02
Design experiment,
develop and implement
data collection process
SDSBA02
Apply Business
Processes
Management (BPM),
general business
processes and
operations for
EDSF Release 2: Part 1. Data Science Competence Framework (CF-DS)
Table 4.3. Required skills related to analytics languages, tools, platforms and Big Data infrastructure 6
DSDALANG
Data Analytics
and Statistical
languages and
tools
DSADB
Databases and
query
languages
DSVIZ
Data/Applicatio
ns visualization
DSADM
Data
Management
and Curation
platform
DSBDA
Big Data
Analytics
platforms
DSDEV
Development and
project
management
frameworks,
platforms and tool
DSDALANG01
R and data analytics
libraries (cran,
ggplot2, dplyr,
reshap2, etc.)
DSADB01
SQL and
relational
databases (open
source:
PostgreSQL,
mySQL, Nettezza,
etc.)
DSVIZ01
Data visualization
Libraries
(mathpoltlib,
seaborn, D3.js,
FusionCharts,
Chart.js, other)
DSADM01
Data modelling
and related
technologies (ETL,
OLAP, OLTP, etc.)
DSBDA01
Big Data and
distributed
computing tools
(Spark,
MapReduce,
Hadoop, Mahout,
Lucene, NLTK,
Pregel, etc.)
DSDEV01
Frameworks: Python,
Java or C/C++, AJAX
(Asynchronous
Javascript and XML),
D3.js (Data-Driven
Documents), jQuery,
others
Source: EU EDISON Data Science Framework (2017)
14
16. A 9-step model for effective use
of Softer Skills in Data Science
Question
Data Analysis Insight
Planning & Design
Presentation &
Distribution
Solution
Buy - in Sign - off
“Contracting” translating
business questions into
actionable, analytical
terms
“Delivering”
expressing analysis &
insight in actionable
business terms
Addressing business
need
Transparency of
activity
Engagement with key
stakeholders
16
18. All this means that Data Scientists
need freedom to work differently
18
https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/lifecycle
19. If organisation is conducive this
can enable new ways of working
19
Day One: Empathise
Day Two: Define
Day Three: Ideate
Day Four: Prototype
Day Five: Test
22. Data Science is being applied
to a wide variety of problems…
22
C2 General
Whether you’re a Retailer, Real Estate, CPG
or Tourism organization, Vodafone Analytics
allows you to use anonymized mobile data
insights to drive decisions on:
● Site Planning
● Geomarketing
● Performance Management
● Investment Analysis
● Smart tourism
New digital
data products
Vodafone Analytics – Location Intelligence
23. Data Science is being applied
to a wide variety of problems…
23
What is attribution modelling?
Each marketing channel
has an effect on the
customer.
Attribution modelling is
to determine these
effects.
https://github.com/5ri/WiML2017/blob/master/Skyscanner-MultiTouchAttribution-Poster.pdf
24. Data Science is being applied
to a wide variety of problems…
24
12
Proactively monitor key performance metrics across all timeseries and dimensions,
autonomously automating the generation of insights
LIBERTY GLOBAL – PERFORMANCE MANAGEMENT
DATA LEADERS SUMMIT | October 18, 2018 | appli Performance Management Application UI Design
Retail Store Gross Additions
Flanders Retail Locations
West Flanders Weekly Retail Gross Additions
Retail Store Weekly Sales per Store
Retail Net New ARPU
109 119 120
Retail gross additions are declining…
16.7
Sample Data
‒ Disguised ‒
Copyright 2018 appli. Inc. Patents pending (USPTO #62562910, #62625645).
All rights reserved.
SAMPLE USE CASES
PERFORMANCE MANAGEMENTPERFORMANCE MANAGEMENT MARKETING ANALYTICSMARKETING ANALYTICS PREDICTIVE MODELINGPREDICTIVE MODELING
Single Guest Vie
Propensity
Modeling
Cost of
Acquisition
A/B
Testing
Return on
Marketing
Investment
$0
$50
$100
$150
$200
$250
$300
0 2,000 4,000 6,000 8,000 10,000
Demand Modeling
Forecasting and Pacing
Sales, Base and Churn Monitoring
• Proactively monitor key performance
metrics across all timeseries and
dimensions, autonomously detecting
anomalies against expectations
Insights Generation
• Automate the production of business
insights; providing visibility into why
performance is anomalous and identify,
size the drivers of influence
Customer Analytics
• Integrate disparate data, both internal
and third-party, to understand
customer behaviors and signals
influencing the customer journey
Marketing Effectiveness
• Determine where to spend the next
dollar of marketing investment by
understanding probabilistic
relationships of marketing levers
Pricing – Demand Modeling
• Model demand and revenue curves to
optimize promotional price offerings
and develop pricing strategies
Forecasting / Pacing
• Utilize business drivers – marketing
mix, network, competitive, seasonal
and more – to generate more reliable
predictions of business outcomes
25. Some European Insurer examples
shared at public conferences
Despite traditional firms generally being
laggards with regards to Data Science &
technology.
Here are a number of examples from other
insurers (shared publicly at conferences):
❖ Vitality (segments + behavioural analytics)
❖ Lloyds Bank (retention analytics/models)
❖ Royal London (predictive underwriting)
❖ Scottish Widows (marketing metrics)
❖ Groupama (data + telematics innovation)
❖ AIG (behavioural biases tests)
❖ Agila (digital analytics + personalisation)
❖ Aviva (intermediary + digital analytics)
25
26. Increasingly the focus is on
mobile-first design principles
NOTIFY A CUSTOMER WHEN
SOMETHING HAPPENS
you need to know what
mattersto a customer, their
latest status and the best
way to alert them, given
the context
RESPOND TO A REQUEST
FOR INFORMATION
you need to know the experience
of a customer and what solutions
you might be able to offer
SHARE AN
EXPERIENCE
to enable social media
sharing you need up-to-
date contact data and all
relevant permissions
COMPLETE A
TRANSACTION
which needs to be
quick and easy, its progress
remembered, and cross-
device migration enabled
COLLABORATE
WITH A CUSTOMER
OR EMPLOYEE
you need to know their
availability, past conversations
and preferences
ENTERTAIN OR
EDUCATE SOMEONE
you need to remember what they
have seen previously in order to
learn about their preferences
CAPTURE OR
CREATE CONTENT
if you enable use of video and/or audio
recording, you need to be able to manage
those ‘Big Data’ challenges
SEVEN TYPES OF DYNAMIC CUSTOMER ENGAGEMENT
However, this will go horribly wrong without accurate dynamic customer data. ‘The Mobile
Mind Shift’4
challenges businesses to design differently for ‘mobile moments’. The authors,
who are Forrester analysts, identify seven types of engagement. Consider how each of
these could go wrong without accurate dynamic customer data:
Source: The Mobile Mind Shift4
With accurate, up-to-date dynamic customer data, organisations
can redesign their processes, thereby meeting increased consumer
expectations of such helpful and engaging experiences.
Could that be a key opportunity for your business?
RESPOND TO A REQU
FOR INFORMATION
you need to know the ex
of a customer and what s
you might be able to offe
COMPLETE A
TRANSACTION
which needs to be
quick and easy, its progress
remembered, and cross-
device migration enabled
CAPTURE OR
CREATE CONTENT
if you enable use of video and/or audio
recording, you need to be able to manage
those ‘Big Data’ challenges
4
The Mobile Mind Shift: Engineer Your Business to Win in the Mobile Moment” by Ted Schadler, Josh Bernoff and Julie Ask (Groundswell Press, 2014)
Source: The Mobile Mind Shift4
26
27. Following some of the leading
Data Science leaders can help
27
Orlando Machado = Aviva Sanjeevan Bala= Channel 4 Lester Berry = John Lewis
Ryan Den Rooijen = Dyson Graeme McDermott = Addison Lee Martin Squires = Homeserve
31. • Video (e.g. accident or property
damage)
• Audio (e.g. customer call)
• Image (e.g. specified items to insure)
• IoT data (e.g. wearable utilisation)
• Text (e.g. emails/text/social media
interaction)
• Models/Rules (e.g. propensity score
or segment for personalisation)
• Event triggers (e.g. Proxies for
renewal dates, important life events)
Digital Transformation is driving a
wider range of data types to handle
31
33. Data is still foundational to Data
Science applications
Monica Rogati: The AI Pyramid of Needs. Hacker-noon blog post, 1 Aug 2017
(https://hacker-noon.com/the-ai-hierarchy-of-needs-18f111fcc007, accessed 19/04/2018).33
34. Old data foundations still matter,
don’t overlook the basics
Predictive + Prescriptive Analytics
Business Intelligence
Data Science & AI
Traditional Analytics
Data Availability & Quality are the foundations for building any part of this house
• Ease of data access (normally through flexible Cloud based solution, AWS et al)
• Freedom to move and transform data (with suitable Data Lake or “sand pit” space available)
• Single Customer View (at least virtual, to enable customer records as basis for analysis
34
35. Latest EU Data Protection Regulation
has raised the bar (GDPR inc. Article 5)
PROTECTING CONSUMERS
The GDPR has been mentioned a few times already, and we all need to consider its
principle of privacy by design, and by default, in our plans. That means ensuring all uses
of new technology, or major changes, consider their impact on data subjects. Normally
this will be via formal Data Protection Impact Assessments (DPIAs). These must include
due consideration being given to ensuring data accuracy and appropriate fair processing.
2017 research by RMDS confirmed that, for marketers,
GPDR non-compliance had reached the top of customer
data management challenges.
CUSTOMER DATA MANAGEMENT CHALLENGES
2017 – All 2017 – Brands 2017 – Agencies 2016
NON COMPLIANCE
WITH GDPR
29.4%
11.9%
24.6%
35.2%
LEGACY
SYSTEMS
27.7%
37.6%
37.3%
16.2%
POOR DATA
EQUALITY
17.8%
20.3%
16.7%
19.1%
SOURCING
RESPONSIVE DATA
10.8%
11.9%
9.5%
12.4%
NO CHALLENGES
4.8%
9.4%
2.4%
7.6%
DON’T KNOW
4.0%
2.6%
0.8%
4.8%
OTHER
6.9%
4.9%
8.7%
4.8%
Source: Royal Mail Data Services Research 2017
DATA ACCURACY CHALLENGE 2:
THE GDPR AND BEYOND
Since May 2018, the General Data Protection Regulation (GDPR) has been in force in the
UK, enforced by the Information Commissioner’s Office (ICO). UK and European businesses
are coming to terms with what this means. Fines of up to 4% of global annual turnover are
possible under GDPR and recent actions by the ICO indicate a growing willingness to act
where an organisation breaches data protection law.
Consumers have also been woken up to their rights
by a flood of privacy policy and repermissioning emails.
So the risk to brand reputation of being ‘called out’ by
the ICO may be the greatest cost. The ‘Your Data Matters’
campaign by the ICO is also raising public awareness.
A great deal of press coverage on the GDPR has focussed on the higher standard for
evidence of ‘consent’ for data processing. Article 6 of the GDPR, outlining potential legal
bases for data processing, does set a higher bar. Whichever bases for data processing are
them, on average, six per cent of their annual revenues.
DON’T KNOW
40%
30%
20%
10%
0%
5.1%
3.7%
3.3%
5.7%
33.2%
33.7%
20.1%
34.6%
30.6%
23.3%
4.2%
2.6%
16+% 11-15% 6-10% <5% 0%
COST OF POOR-QUALITY CUSTOMER CONTACT DATA AS A PERCENTAGE OF ANNUAL REVENUE
Source: Royal Mail Data Services Research 2017
CONCLUSION
In conclusion, dynamic customer data is both a key challenge for today’s businesses and
a potential win-win benefit. Either way, it cannot be ignored. Accurate data has become
the lifeblood of today’s business operations and customer interactions. The need for
timely and accurate dynamic customer data management has never been greater.
‘Dynamic customer data’ is a term reflecting how quickly customer
data is changing, leading to out-of-date data in businesses.
However, there is a positive business case to be made for investing in a solution to
achieve dynamic customer data management. ROI benefits may come from a wide
range of functions including marketing cost savings, improved customer experience
and compliance fine avoidance.
LEGISLATION
(GDPR)
SPEED OF
CHANGE
DEMANDS
ON IT
THREE KEY CHALLENGES
FOR TODAY’S BUSINESSES MAKE ACHIEVING
DATA ACCURACY HARDER:
THREE KEY TRENDS
PROVIDE EXAMPLES OF CUSTOMER ACQUISITION AND
RETENTION OPPORTUNITIES, FROM USE OF DATA:
35
36. The digitisation of our lives also drives
a faster speed of Data Quality erosion
DATA ACCURACY CHALLENGE 2:
Research from Royal Mail Data Services (RMDS) reveals that
organisations believe that inaccurate customer data costs
them, on average, six per cent of their annual revenues.
DON’T KNOW
40%
30%
20%
10%
0%
5.1%
3.7%
3.3%
5.7%
33.2%
33.7%
20.1%
34.6%
30.6%
23.3%
4.2%
2.6%
16+% 11-15% 6-10% <5% 0%
COST OF POOR-QUALITY CUSTOMER CONTACT DATA AS A PERCENTAGE OF ANNUAL REVENUE
Source: Royal Mail Data Services Research 2017
complex interconnected digital economy, that the GiGo theory still applies – that is to say
‘garbage in = garbage out’.
A STEEP CLIMB OVER DATA ACCURACY CHALLENGES
Data gurus3
have offered advice on customer data quality management for over 20 years.
However, many factors make this a more complex challenge than those faced in the past.
The speed of changes in personal data, increased data protection regulation and the
amount of IT change planned all make this feel like a moving target.
Let’s consider those challenges in turn, to see how they may apply to your business...
DATA ACCURACY CHALLENGE 1: ‘SPEED OF CHANGE’
Changes in consumer behaviour, as well as a greater need for up-to-date data, mean
customer data decays at a faster rate than ever before. Akin to use-by dates on food,
data becomes useless if it is not updated to keep track of changing circumstances.
This rate of personal data change has gone unnoticed by too many organisations.
Looking at this in more detail, up to 3,000 changes are made every day to the Royal Mail
Postcode Address File (PAF®
). Couple this with the latest data from the Office of National
Statistics relating to daily life events that also affect customer data accuracy, there are so
many changes occurring on a daily basis for this to easily and quickly lead to out-of-date
addresses, names or inappropriate understanding of needs. The hidden cost to your
business of such inaccurate data reveals itself in many business functions. Inaccurate
addresses cause marketing and product delivery returns, as well as impacting billing
and collections. Inaccurate names deliver poor customer experience, and customers
may leave due to a poor impression of the company. In addition to marketing and
operational costs, there is a greater risk of regulatory fines.
9,590HOUSEHOLDS MOVE
1,496PEOPLE MARRY
810PEOPLE DIVORCE
2,011PEOPLE RETIRE
1,500PEOPLE DIE
DAILY LIFE EVENTS
AFFECTING CUSTOMER
DATA ACCURACY
Source: Office of National Statistics
36
37. The majority of businesses are not
cleaning their data often enough
pace. More than 15,000 changes are being made daily to people’s personal information,
which equates to 5.6 million changes per year.
The term ‘dynamic customer data’ has been coined to remind us of this complexity. In this
context, data accuracy management is not a ‘once and done’ activity. Organisations cannot
rely on cleaning data on import to a data warehouse, just to then leave the data to rot in
that ‘data graveyard’.
The regularity of data changes, and the need to have up-to-date accurate customer data,
requires rapid data cleaning. Dynamic customer data management is a term covering the
technologies that provide this. These technologies enable organisations to clean new data
and update existing data daily or in near real-time. They allow businesses to keep up with
the consumer life events affecting their customer data.
Worryingly, research from RMDS reveals that almost a third of businesses have no formal
data cleaning process; less than a quarter of the businesses surveyed are implementing
data cleaning on a daily or continuous basis.
FREQUENCY OF DATA CLEANING
DAILY/
CONTINUOUSLY
NO FORMAL
PROCESS
DON’T KNOWMONTHLY QUARTERLY ANNUALLY
40%
30%
20%
10%
0%
18.7%
13.6% 13.1%
9.8% 10.8%
14%
32.7%
37.3%
7.9%
12.4%
7.8%
22%
Source: Royal Mail Data Services Research 2017
2016 2017
to increase the need for positive consent.
Beyond data protection,
a number of sectors are
also being challenged
to better safeguard
their customers,
from conduct risk for financial services firms to
increased scrutiny for gambling businesses. As
an example of dynamic customer data impact,
one gambling brand discovered more than 75
duplicate records for one customer. That scale
of error makes safeguarding individuals or
conducting risk management impossible.
Data accuracy is essential for operational processes to protect individuals.
It is also needed for executive reporting and as evidence for regulators.
Failing to address corrupted data accuracy can lead to flawed decisions,
some of which may be critical to keeping your customers.
37
38. Requirements for a Dynamic
Customer Data Management solution
38
DYNAMIC CUSTOMER DATA
MANAGEMENT SOLUTIONS
REQUIREMENT 1: QUALITY OF REFERENCE DATA
If you are feeling under-prepared, and without a solution to move from your current legacy
systems to this brave new day, there is hope.
Reassuringly, technology solutions are available to meet this need. But, like so much IT,
not all of them are as complete as they appear. So, let’s briefly consider what to look for
in such a solution.
Using external data sources to validate, correct or augment internal customer
data relies on the quality of sources. Check the coverage, quality control,
timeliness and ownership of original data sources. Beware of an impressive
demo, or initial data set, that will degrade over time because of provenance.
Keeping up with dynamic customer data often requires a provider who
owns key data assets. Your due diligence should also require evidence of
their GDPR compliance – for example, the precise permission gained or
how data subjects are kept informed.
REQUIREMENT 2: EASE OF ENTERPRISE-WIDE INTEGRATION
For most businesses, beset with a complex legacy infrastructure, more
on-site systems are not the solution. Look for cloud-based solutions,
combined with easy to use application programming interfaces (APIs).
Such a solution should achieve a lower cost of ownership. It will also
be more flexible to integrate with changing internal systems. Involve
your IT team, and ask providers for details of APIs, as well as any push
notifications from external systems.
REQUIREMENT 3: SECURE AND COMPLIANT SHARING OF DATA
Given one of the key reasons for investing in such a solution is to comply
with the GDPR, this should be a focus. There is no point improving your
compliance with regards to data accuracy, only to breach rules on data
sharing. Look for evidence of security and anonymity. What levels of
DYNAMIC CUSTOMER DATA
MANAGEMENT SOLUTIONS
REQUIREMENT 1: QUALITY OF REFERENCE DATA
If you are feeling under-prepared, and without a solution to move from your current legacy
systems to this brave new day, there is hope.
Reassuringly, technology solutions are available to meet this need. But, like so much IT,
not all of them are as complete as they appear. So, let’s briefly consider what to look for
in such a solution.
Using external data sources to validate, correct or augment internal customer
data relies on the quality of sources. Check the coverage, quality control,
timeliness and ownership of original data sources. Beware of an impressive
demo, or initial data set, that will degrade over time because of provenance.
Keeping up with dynamic customer data often requires a provider who
owns key data assets. Your due diligence should also require evidence of
their GDPR compliance – for example, the precise permission gained or
how data subjects are kept informed.
REQUIREMENT 2: EASE OF ENTERPRISE-WIDE INTEGRATION
For most businesses, beset with a complex legacy infrastructure, more
on-site systems are not the solution. Look for cloud-based solutions,
combined with easy to use application programming interfaces (APIs).
Such a solution should achieve a lower cost of ownership. It will also
be more flexible to integrate with changing internal systems. Involve
your IT team, and ask providers for details of APIs, as well as any push
notifications from external systems.
REQUIREMENT 3: SECURE AND COMPLIANT SHARING OF DATA
Given one of the key reasons for investing in such a solution is to comply
with the GDPR, this should be a focus. There is no point improving your
compliance with regards to data accuracy, only to breach rules on data
sharing. Look for evidence of security and anonymity. What levels of
encryption are supported for data sent to or from a cloud-based solution?
After initial data provision, how are updates kept anonymous to avoid data
breaches? How secure are their systems and premises?
REQUIREMENT 1: QUALITY OF REFERENCE DATA
If you are feeling under-prepared, and without a solution to move from your current legacy
systems to this brave new day, there is hope.
Reassuringly, technology solutions are available to meet this need. But, like so much IT,
not all of them are as complete as they appear. So, let’s briefly consider what to look for
in such a solution.
Using external data sources to validate, correct or augment internal customer
data relies on the quality of sources. Check the coverage, quality control,
timeliness and ownership of original data sources. Beware of an impressive
demo, or initial data set, that will degrade over time because of provenance.
Keeping up with dynamic customer data often requires a provider who
owns key data assets. Your due diligence should also require evidence of
their GDPR compliance – for example, the precise permission gained or
how data subjects are kept informed.
REQUIREMENT 2: EASE OF ENTERPRISE-WIDE INTEGRATION
For most businesses, beset with a complex legacy infrastructure, more
on-site systems are not the solution. Look for cloud-based solutions,
combined with easy to use application programming interfaces (APIs).
Such a solution should achieve a lower cost of ownership. It will also
be more flexible to integrate with changing internal systems. Involve
your IT team, and ask providers for details of APIs, as well as any push
notifications from external systems.
REQUIREMENT 3: SECURE AND COMPLIANT SHARING OF DATA
Given one of the key reasons for investing in such a solution is to comply
with the GDPR, this should be a focus. There is no point improving your
compliance with regards to data accuracy, only to breach rules on data
sharing. Look for evidence of security and anonymity. What levels of
encryption are supported for data sent to or from a cloud-based solution?
After initial data provision, how are updates kept anonymous to avoid data
breaches? How secure are their systems and premises?
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
To avoid regulatory fines and take appropriate customer action, achieving
a single version of the truth is essential. Look for solutions capable of
maintaining as near to a real-time single customer view. This is likely to
require several elements of functionality from potential suppliers:
REQUIREMENT 4: REAL-TIME ‘SINGLE CUSTOMER VIEW’ (SCV)
• An initial complete cleanse of all customer data
• Generation, maintenance and sharing of encrypted, unique customer keys
• Integration with all systems across your organisation, to review all new transactions
• Ability to provide these capabilities on demand ‘as a service’
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
REQUIREMENT 6: A PARTNER YOU CAN TRUST
Just as Rome wasn’t built in a day, so this will not be a ‘once and done’
type solution. Achieving, maintaining and improving the quality of your
customer data is a long-term commitment. Your business will continue
to need such data accuracy for years to come. So, as with all strategic
IT investment, the people matter as much as the technology. Look for a
provider you can trust, one who understands the business of customer
data with an obvious track record of achieving data accuracy and
managing the issues that arise.
Ensure too that they not only understand the implications of the GDPR,
but that they are people you want to work with. A provider who invests
in events and content to educate others is a good sign.
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
20
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
REQUIREMENT 6: A PARTNER YOU CAN TRUST
Just as Rome wasn’t built in a day, so this will not be a ‘once and done’
type solution. Achieving, maintaining and improving the quality of your
customer data is a long-term commitment. Your business will continue
to need such data accuracy for years to come. So, as with all strategic
IT investment, the people matter as much as the technology. Look for a
provider you can trust, one who understands the business of customer
data with an obvious track record of achieving data accuracy and
managing the issues that arise.
Ensure too that they not only understand the implications of the GDPR,
but that they are people you want to work with. A provider who invests
in events and content to educate others is a good sign.
maintaining as near to a real-time single customer view. This is likely to
require several elements of functionality from potential suppliers:
• An initial complete cleanse of all customer data
• Generation, maintenance and sharing of encrypted, unique customer keys
• Integration with all systems across your organisation, to review all new transactions
• Ability to provide these capabilities on demand ‘as a service’
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
REQUIREMENT 6: A PARTNER YOU CAN TRUST
Just as Rome wasn’t built in a day, so this will not be a ‘once and done’
type solution. Achieving, maintaining and improving the quality of your
customer data is a long-term commitment. Your business will continue
to need such data accuracy for years to come. So, as with all strategic
IT investment, the people matter as much as the technology. Look for a
provider you can trust, one who understands the business of customer
data with an obvious track record of achieving data accuracy and
managing the issues that arise.
To avoid regulatory fines and take appropriate customer action, achieving
a single version of the truth is essential. Look for solutions capable of
maintaining as near to a real-time single customer view. This is likely to
require several elements of functionality from potential suppliers:
REQUIREMENT 4: REAL-TIME ‘SINGLE CUSTOMER VIEW’ (SCV)
• An initial complete cleanse of all customer data
• Generation, maintenance and sharing of encrypted, unique customer keys
• Integration with all systems across your organisation, to review all new transactions
• Ability to provide these capabilities on demand ‘as a service’
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
REQUIREMENT 6: A PARTNER YOU CAN TRUST
Just as Rome wasn’t built in a day, so this will not be a ‘once and done’
type solution. Achieving, maintaining and improving the quality of your
customer data is a long-term commitment. Your business will continue
to need such data accuracy for years to come. So, as with all strategic
IT investment, the people matter as much as the technology. Look for a
provider you can trust, one who understands the business of customer
data with an obvious track record of achieving data accuracy and
managing the issues that arise.
Ensure too that they not only understand the implications of the GDPR,
To avoid regulatory fines and take appropriate customer action, achieving
a single version of the truth is essential. Look for solutions capable of
maintaining as near to a real-time single customer view. This is likely to
require several elements of functionality from potential suppliers:
REQUIREMENT 4: REAL-TIME ‘SINGLE CUSTOMER VIEW’ (SCV)
• An initial complete cleanse of all customer data
• Generation, maintenance and sharing of encrypted, unique customer keys
• Integration with all systems across your organisation, to review all new transactions
• Ability to provide these capabilities on demand ‘as a service’
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
REQUIREMENT 5: INTELLIGENCE OF CUSTOMER MATCHING RULES
Historically, efforts to achieve SCVs have been hampered by simplistic
matching. Rules or algorithms have not managed to handle the subtle
variations people make to name and address spelling. Look for a
provider who has experience of handling such nuances in addressing.
Ask them about how their rules would handle all the different ways
people address you.
REQUIREMENT 6: A PARTNER YOU CAN TRUST
Just as Rome wasn’t built in a day, so this will not be a ‘once and done’
type solution. Achieving, maintaining and improving the quality of your
customer data is a long-term commitment. Your business will continue
to need such data accuracy for years to come. So, as with all strategic
IT investment, the people matter as much as the technology. Look for a
provider you can trust, one who understands the business of customer
data with an obvious track record of achieving data accuracy and
managing the issues that arise.
Ensure too that they not only understand the implications of the GDPR,
To avoid regulatory fines and take appropriate customer action, achieving
a single version of the truth is essential. Look for solutions capable of
maintaining as near to a real-time single customer view. This is likely to
require several elements of functionality from potential suppliers:
REQUIREMENT 4: REAL-TIME ‘SINGLE CUSTOMER VIEW’ (SCV)
• An initial complete cleanse of all customer data
• Generation, maintenance and sharing of encrypted, unique customer keys
• Integration with all systems across your organisation, to review all new transactions
• Ability to provide these capabilities on demand ‘as a service’
• Encrypted push notifications of changes to be made to related customer records
• Near-real-time updates, notifying key changes in customer data
39. Pause to Think:
What data is currently not
being used (or cleaned)
in your organisation?
48. 5 keys from experience of helping
businesses implement Data Science
48
1. Access to Data
2. Right Tools for the job
3. Domain knowledge available
4. IT team in the loop
5. Clarity of goals/priorities
50. Don’t be confused by different
software labels
50
the industry, it is vital that insurers quickly start to accelerate the acquisition of value from their analytics program.
BUSINESS CAPABILITIES FOR INSURERS
Answers Insurers Are Seeking
Business intelligence and analytics are becoming increasingly vital to every part of the insurance business. Insurers
need capabilities that address a wide variety of questions across marketing, sales, and service as well as enterprise
operations. The general types of questions raised are illustrated in Figure 1. At a high level, insurers want to explore
questions like the ones above the orange boxes: How do we gain new insights from historical data? What are our
new opportunities? At the next level, they are asking more specific questions, like those inside the orange boxes:
What happened? Why is it happening? What can we do about it?
There are a variety of technology tools and approaches to address these questions. They generally fall under the
categories of business intelligence, advanced analytics, and emerging analytics, and these technologies can be
applied to answer the types of questions posed in each of the sections.
Figure 1. The BI and Analytics Spectrum for Insurers
Source: Strategy Meets Action 2016
As can be seen from the diagram, the questions to be answered range from the traditional, more operational
types of issues to more complex and differentiating insights and actions. On the far left, analysis of historical
data enables reporting on the state of the business (What happened? What is happening now? Where is the
problem?). In addition, historical data (both internal and external) can help with diagnostics on specific problems
(Why is it happening? What if it continues?). Towards the middle and the right of the diagram, more complex
and forward-looking analytics can be applied to understand how insurers can identify, predict, and capitalize on
new opportunities, and ultimately, through emerging analytics, move to human augmentation and automated
decisioning. It should be noted that big data is an overlay onto this diagram, providing a set of approaches and
technologies to answer these questions when the volume, variety, and velocity of the data cannot be addressed in
a timely manner by traditional analytics.
How do we gain new insights
from historical data?
BUSINESS INTELLIGENCE
What are our new
opportunities?
How do we capitalize on new
opportunities?
How do we leverage
human intelligence?
ADVANCED
ANALYTICS
EMERGING
ANALYTICS
Analytic
Collab-
oration
Predictive
Modules
Predictive
Analytics
Data & Text
Mining
Advanced
Statistical
Analysis
ScenariosAnalysis
Ad-hoc
Queries
Dashboards
&
Scorecards
Reporting
Cognitive
Computing
What can we do about it?What is likely to happen?
What if it
continues?
Why is it happening?
Where
is the
problem?
What is
happening?
What
happened?
51. The critical need to manage
domain knowledge
51
Tips to develop domain knowledge ready
for your Data Scientists:
1. “Know your numbers” = develop
commercial awareness in all analysts
2. Use team meetings to ensure clarity on
big picture & the “why” of work
3. Get out & about, including potentially
shadowing or job swaps (e.g. Strategy)
52. Need to manage relationship with
IT carefully (like a marriage)
52
The best home for Data Science?
Benefits of “end user computing”, but the
need for process to “promote to live”
Centralised verses Decentralised teams?
Central Design Authority?
Greater data access & the need for a
“playpen” or “data lake” for experiments
Data and Goliath
If de-centralised, where does your
data function and team members sit?
These interesting results
reflect both a significant
change from last year and
a concern I have about IT
ownership. Last year we
saw that most decentralised
teams were scattered across
businesses (46%), which can be
very inefficient.
However, the picture has now
changed to most decentralised
data teams residing in parts of
IT. That concerns me. When
data governance or analytics
are left to IT, business leaders
can tend to obfuscate their
responsibilities , and analytics
can be reduced to inflexible
projects.
I have seen much more
success with data expertise
residing within business and
with analytics functions
sitting within Marketing or
Operations. I hope this is not a
further step on the road where
Data Science is in danger of
reducing analytics to coding.
ឣ IT 51%
ឣ Scattered across the business 27%
ឣ Marketing 15%
ឣ Operations 7%
51% 27% 15% 7%
As a result of
decentralisation, 51%
of our respondents’
data teams are now
located within the IT
department.
Key Findings:
53. How to achieve clarity of goals &
why that is needed
53
Don’t hire a Data Scientist without
being clear on how they could help
Don’t hire a Data Scientist without
being ready to set them clear goals
Do work with them to prioritise
business challenges & opportunities
54. Pause to Think:
Are you ready to start
using Data Science?
If not, what are your gaps?
55. Action-orientated education:
the biggest predictor of value…
55
One thing I will do
differently as a result of
today is…
01 All around us, but let’s get clearer
What is it?
02 Some exciting & concerning applications
How is it being used?
04 How could you learn the coding needed?
Coding Opportunities
03 How can you spot new data & uses?
Data Opportunities
05 What is needed to get started?
Getting ready
56. Data Science is about a learning
journey, here’s a blog to help you
56
57. Other resources to help your CPD
EDISON Data Science Framework:
Part 1. Data Science Competence Framework (CF-DS)
Release 2
Project acronym: EDISON
Project full title: Education for Data Intensive Science to Open New science frontiers
Grant agreement no.: 675419
Due Date
Actual Date 3 July 2017
Document Author/s Yuri Demchenko, Adam Belloum, Tomasz Wiktorski
Version Release 2, v0.8
Dissemination level PU
Status Working document, request for comments
Document approved by
This work is licensed under the Creative Commons Attribution 4.0 International License.
To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
57
58. Checking in on Baron von
Munchhausen…
58
Goals:
Did you get what you
needed from this
masterclass?