Technology Office challenges especially better use of information within a banking environment. Provided tangible examples of transactional analysis, text sentiment analysis, power consumption analysis
Credit Card Fraud Detection Using ML In DatabricksDatabricks
In the Credit Card Companies, illegitimate credit card usage is a serious problem which results in a need to accurately detect fraudulent transactions vs non-fraudulent transactions. All organizations can be hugely impacted by fraud and fraudulent activities, especially those in financial services. The threat can originate from internal or external, but the effects can be devastating – including loss of consumer confidence, incarceration for those involved, even up to downfall of a corporation. Despite regular fraud prevention measures, these are constantly being put to the test in an attempt to beat the system.
Fraud detection is a task of predicting whether a card has been used by the cardholder. One of the methods to recognize fraud card usage is to leverage Machine Learning (ML) models. In order to more dynamically detect fraudulent transactions, one can train ML models on a set of dataset including credit card transaction information as well as card and demographic information of the owner of the account. This will be our goal of the project while leveraging Databricks.
Business intelligence data analytics-visualizationMuthu Natarajan
Business Intelligence, Cloud Computing, Data Analytics, Data Scrubbing, Data Mining, Big Data & Intelligence, How to use Data into Information, Decision Based, Methods for Business Intelligence, Advanced Analytics, OLAP, Multidimensional Data, Data Visualization.
This document discusses opportunities for using big data in private wealth management. It begins by defining big data and describing how data volumes have increased exponentially. It then outlines several potential use cases for big data in areas like real-time performance metrics, portfolio optimization, and leveraging customer data. For each use case, it describes current limitations and how a big data approach could enable new capabilities. Finally, it proposes a phased approach for wealth managers to identify use cases, prioritize them, implement proofs of concept, and incrementally automate analysis and reporting. The overall message is that big data can enhance analytics and open up new opportunities previously only available to investment banks.
Basic Concepts of Business Data Analytics, Evolution of Business Analytics, Data Analytics, Business Data Analytics Applications, Scope of Business Analytics.
The document discusses business analytics and the role of a business analyst. It defines key terms like business analytics, data analytics, business intelligence, big data, data science, and data mining. It describes the skills required of a business analyst like understanding the business, basic statistics, Excel, and some analytics tools. The duties of a business analyst are to understand business problems and use data to help decision making. The document also lists some common business analyst job titles and roles.
Data science involves using industrial research techniques on a company's own data to develop advanced algorithms that provide a competitive advantage. Data engineering is a specialized form of software engineering focused on handling and processing data using skills in areas like structured and unstructured data storage, machine learning platforms, and predictive APIs. While data science and business intelligence overlap in using data analysis, statistics, and visualization, data science has a more scientific approach focused on the future rather than the past. Data-focused jobs are in high demand across many industries, especially technology, but some roles may become automated, increasing the value of skills like research and communication. Education options for these fields include academic programs, boot camps, and online classes.
This document discusses the role of a Chief Analytics Officer within the Science and Technology Directorate (S&T) of the Department of Homeland Security (DHS). The S&T mission is to deliver solutions to critical homeland security needs by conducting research, development, testing and evaluation. The proposed role of the Chief Analytics Officer is to develop and execute analytic strategies to improve decision making within S&T and DHS through independent analysis, data-driven insights, and anticipatory analytics capabilities. Initial responsibilities would include portfolio analysis, horizon scanning, and prototyping anticipatory analytics with DHS components.
Credit Card Fraud Detection Using ML In DatabricksDatabricks
In the Credit Card Companies, illegitimate credit card usage is a serious problem which results in a need to accurately detect fraudulent transactions vs non-fraudulent transactions. All organizations can be hugely impacted by fraud and fraudulent activities, especially those in financial services. The threat can originate from internal or external, but the effects can be devastating – including loss of consumer confidence, incarceration for those involved, even up to downfall of a corporation. Despite regular fraud prevention measures, these are constantly being put to the test in an attempt to beat the system.
Fraud detection is a task of predicting whether a card has been used by the cardholder. One of the methods to recognize fraud card usage is to leverage Machine Learning (ML) models. In order to more dynamically detect fraudulent transactions, one can train ML models on a set of dataset including credit card transaction information as well as card and demographic information of the owner of the account. This will be our goal of the project while leveraging Databricks.
Business intelligence data analytics-visualizationMuthu Natarajan
Business Intelligence, Cloud Computing, Data Analytics, Data Scrubbing, Data Mining, Big Data & Intelligence, How to use Data into Information, Decision Based, Methods for Business Intelligence, Advanced Analytics, OLAP, Multidimensional Data, Data Visualization.
This document discusses opportunities for using big data in private wealth management. It begins by defining big data and describing how data volumes have increased exponentially. It then outlines several potential use cases for big data in areas like real-time performance metrics, portfolio optimization, and leveraging customer data. For each use case, it describes current limitations and how a big data approach could enable new capabilities. Finally, it proposes a phased approach for wealth managers to identify use cases, prioritize them, implement proofs of concept, and incrementally automate analysis and reporting. The overall message is that big data can enhance analytics and open up new opportunities previously only available to investment banks.
Basic Concepts of Business Data Analytics, Evolution of Business Analytics, Data Analytics, Business Data Analytics Applications, Scope of Business Analytics.
The document discusses business analytics and the role of a business analyst. It defines key terms like business analytics, data analytics, business intelligence, big data, data science, and data mining. It describes the skills required of a business analyst like understanding the business, basic statistics, Excel, and some analytics tools. The duties of a business analyst are to understand business problems and use data to help decision making. The document also lists some common business analyst job titles and roles.
Data science involves using industrial research techniques on a company's own data to develop advanced algorithms that provide a competitive advantage. Data engineering is a specialized form of software engineering focused on handling and processing data using skills in areas like structured and unstructured data storage, machine learning platforms, and predictive APIs. While data science and business intelligence overlap in using data analysis, statistics, and visualization, data science has a more scientific approach focused on the future rather than the past. Data-focused jobs are in high demand across many industries, especially technology, but some roles may become automated, increasing the value of skills like research and communication. Education options for these fields include academic programs, boot camps, and online classes.
This document discusses the role of a Chief Analytics Officer within the Science and Technology Directorate (S&T) of the Department of Homeland Security (DHS). The S&T mission is to deliver solutions to critical homeland security needs by conducting research, development, testing and evaluation. The proposed role of the Chief Analytics Officer is to develop and execute analytic strategies to improve decision making within S&T and DHS through independent analysis, data-driven insights, and anticipatory analytics capabilities. Initial responsibilities would include portfolio analysis, horizon scanning, and prototyping anticipatory analytics with DHS components.
Presentation on "A Complete Overview of Data Driven Decision Making in a Quickly Changing Business Environment" given by Isaac Aidoo, Head of Data Analytics, Zoona.
Four Techniques to Run AI on Your Business DataHyoun Park
This webinar explores key topics for preparing AI, including
Is Your Data Ready for AI?
Why Do We Mean By AI?
Practical AI for Business Profit
Preparing Data for AI
From BI to AI
Key Techniques for Running AI on Business Data
Loras College 2016 Business Analytics Symposium KeynoteRich Clayton
Leaders who embrace data have a profound impact on their organizations yet too few seize the opportunity. Biases in decision making, technology myths, data quality and analytical skills and are the most frequently cited obstacles by organizations of all sizes. Technology advances have neutralized the scale advantage and have democratized analytics for every organization – so now what? Are you to engage more data in your management decisions? Do you have an analytic strategy that has two speeds – one for innovation and one for scale? Are you investing in your top talent so they can ask new questions?
We’ll explore these topics and how to create an analytic culture in your organization. We’ll share how leaders have transformed their organizations by innovating their analytic processes, re-designing the way they work and embracing new technology innovation. We’ll dispel myths about technology and provide you a foundation for building your journey to analytic excellence.
Business Intelligence And Business Analytics | ManagementTransweb Global Inc
Business Intelligence is the initial basic step of Business Analytics. It refers to gathering raw and complex data, and converting it into systematic and logical information in a format that is usable by the end user. Copy the link given below and paste it in new browser window to get more information on Business Intelligence And Business Analytics:-
http://www.transtutors.com/homework-help/management/managing-information-technology/business-intelligence-analytics/
This document provides an introduction to predictive analytics. It defines analytics and predictive analytics, comparing their purposes and differences. Analytics uses past data to understand trends while predictive analytics anticipates the future. Business intelligence involves using data to support decision making and aims to provide historical, current and predictive views of business. As technologies advanced, business intelligence evolved from being organized under IT to potentially being aligned under strategy management. Effective communication between business and analytics professionals is important for organizations to benefit from predictive analytics. The business case for predictive analytics includes enabling strategic planning, competitive analysis, and improving business processes to work smarter.
Business analytics workshop presentation finalBrian Beveridge
This document outlines an agenda and presentation for a business analytics seminar for credit union executives and board directors. The presentation will define business analytics, explain how it can help credit unions address key issues like margin compression and regulatory compliance, and provide examples of how analytics can be applied to areas like marketing, risk management, and branch performance. Attendees will learn how predictive analytics can help credit unions retain members, optimize pricing, and streamline operations. The presentation will also cover getting started with business analytics projects.
The document provides an overview of business analytics (BA) including its history, types, examples, challenges, and relationship to data mining. BA involves exploring past business performance data to gain insights and guide planning. It can focus on specific business segments. Types of BA include descriptive analytics like reporting, affinity grouping, and clustering, as well as predictive analytics. Challenges to BA include acquiring high quality data and rapidly processing large volumes of data. Data mining is an important task within BA that helps handle large datasets and specific problems.
Credit card fraud detection using python machine learningSandeep Garg
This document provides an overview of machine learning tools, technologies, and the data preparation process. It discusses collecting and selecting relevant data, data visualization, labeling data for supervised learning, and transforming raw data into a tidy format. The document also covers various data preprocessing techniques, including data cleaning, formatting, handling missing values and outliers, smoothing, aggregation, generalization, and data reduction methods. The goal of these preprocessing steps is to prepare raw data into a structured format suitable for machine learning modeling.
The Business Analytics Value PropositionEric Stephens
Presentation made to the Nashville Technology Council Analytics Peer Network meeting on May 30, 2013. Discussion of the impact of analytics to an organization, along with use cases that can help convey the value of the practice to executives and other managers.
The speaker discusses the importance of evaluating big data analysis to improve projects. They recommend getting a second opinion on methodology or using another data source for verification. A case study on estimating traffic congestion is presented where ground truth sensor data was collected and compared to estimates using metrics to provide feedback. Lessons include ground truth not being easy to obtain, using the right tools like Python and pandas for evaluation, and having an iterative workflow for timely feedback.
System Dynamics, Analytics & Big Data (16th Conference of the UK Chapter of t...Michael Mortenson
The document discusses the relationship between analytics, big data, and system dynamics. It begins by providing background on the growth of analytics and big data. It then discusses relationship problems between analytics and operations research. The main part of the document introduces the Dianoetic Management Paradigm to describe the evolution of management thinking and related technologies over time. It also describes categories of analytics from descriptive to predictive to prescriptive. Finally, it discusses implications of big data for system dynamics, including opportunities around high-volume, unstructured, and streaming data as well as related technologies.
This document provides tips for aspiring data scientists. It advises them to start by focusing on a topic that interests them and to clearly define their objectives and data collection process. It also recommends that they visualize their data, understand the context, look for additional insights, evaluate results, and find effective uses of the data. The document notes that data is becoming increasingly important in all industries and companies without data-savvy managers will be at a disadvantage.
The document discusses business analytics and decision making. It defines key concepts like data warehousing, data mining, business intelligence, descriptive analytics, predictive analytics, and prescriptive analytics. It explains how these concepts are used to extract insights from data to support decision making in organizations. Examples of how different types of analytics can be applied in a retail context are provided.
This presentation introduces big data and explains how to generate actionable insights using analytics techniques. The deck explains general steps involved in a typical analytics project and provides a brief overview of the most commonly used predictive analytics methods and their business applications.
Vijay Adamapure is a Data Science Enthusiast with extensive experience in the field of data mining, predictive modeling and machine learning. He has worked on numerous analytics projects ranging from healthcare, business analytics, renewable energy to IoT.
Vijay presented these slides during the Internet of Everything Meetup event 'Predictive Analytics - An Overview' that took place on Jan. 9, 2015 in Mumbai. To join the Meetup group, register here: http://bit.ly/1A7T0A1
[Ai in finance] AI in regulatory compliance, risk management, and auditingNatalino Busa
AI to Improve Regulatory Compliance, Governance & Auditing. How AI identifies and prevents risks, above and beyond traditional methods. Techniques and analytics that protect customers and firms from cyber-attacks and fraud. Using AI to quickly and efficiently provide evidence for auditing requests.
Data analytics refers to the broad field of using data and tools to make business decisions, while data analysis is a subset that refers to specific actions within the analytics process. Data analysis involves collecting, manipulating, and examining past data to gain insights, while data analytics takes the analyzed data and works with it in a meaningful way to inform business decisions and identify new opportunities. Both are important, with data analysis providing understanding of what happened in the past and data analytics enabling predictions about what will happen in the future.
Material for the 26 Oct 2015 lecture I held for Aalto University business students. The lecture focuses on the high level topics in analytics and Big Data that are either central to the subject or just highly visible in the media.
The main messages of the lecture are:
- The purpose of analytics and of the data analyst is to solve business problems
- Big Data brings over some very special traits to doing analytics that don't exist when working working with smaller datasets. Understanding these traits is a must for successful analytics.
- Deploying analytics is more dependent on humans than on technology
- Data and analytics are nowadays significant assets to many companies. Therefore they need their own strategy and need to be managed just like any other business critical assets.
Learn about the emerging field of big data and advanced quantitative models and how the Rady School's MS in Business Analytics program is designed to solve important business problems.
HYDROLIFT 24, 2003, £49,995 For Sale Brochure. Presented By yachtingelite.comWolfgang Stolle
HYDROLIFT 24
Puerto Banus, Spain
Hyrdolift 24 With Single Mercruiser Inboard 496 Ho Petrol Engines (430hp). This Particular Model Is In Fantastic Condition, Benefitting From A New Paint Job, New Engines (2 Hours Use) And New Outdrives. In Fantastic Condition And Offers Would Be Seriously Considered. Please Contact A Member Of Our Team For Further Information Or Indeed To Arrange A Viewing.
For Sale Brochure. Presented By yachtingelite.com. Visit Site http://www.yachtingelite.com by yachtingelite.com. Real Estate,Luxury Property
как удержать В2В клиентов с помощью изучения целевой аудиторииEkaterina Gould
основываясь на опыте 14ти лет работы в сфере В2В, мы сформулировали эти принципы работы с клиентами. Эмоциональная связь с потребителем, регулярная работа с CRM базой, знание поведенческих характеристик клиента. В нынешних экономических условиях Client is a King! дайте то, что нужно только ему.
Presentation on "A Complete Overview of Data Driven Decision Making in a Quickly Changing Business Environment" given by Isaac Aidoo, Head of Data Analytics, Zoona.
Four Techniques to Run AI on Your Business DataHyoun Park
This webinar explores key topics for preparing AI, including
Is Your Data Ready for AI?
Why Do We Mean By AI?
Practical AI for Business Profit
Preparing Data for AI
From BI to AI
Key Techniques for Running AI on Business Data
Loras College 2016 Business Analytics Symposium KeynoteRich Clayton
Leaders who embrace data have a profound impact on their organizations yet too few seize the opportunity. Biases in decision making, technology myths, data quality and analytical skills and are the most frequently cited obstacles by organizations of all sizes. Technology advances have neutralized the scale advantage and have democratized analytics for every organization – so now what? Are you to engage more data in your management decisions? Do you have an analytic strategy that has two speeds – one for innovation and one for scale? Are you investing in your top talent so they can ask new questions?
We’ll explore these topics and how to create an analytic culture in your organization. We’ll share how leaders have transformed their organizations by innovating their analytic processes, re-designing the way they work and embracing new technology innovation. We’ll dispel myths about technology and provide you a foundation for building your journey to analytic excellence.
Business Intelligence And Business Analytics | ManagementTransweb Global Inc
Business Intelligence is the initial basic step of Business Analytics. It refers to gathering raw and complex data, and converting it into systematic and logical information in a format that is usable by the end user. Copy the link given below and paste it in new browser window to get more information on Business Intelligence And Business Analytics:-
http://www.transtutors.com/homework-help/management/managing-information-technology/business-intelligence-analytics/
This document provides an introduction to predictive analytics. It defines analytics and predictive analytics, comparing their purposes and differences. Analytics uses past data to understand trends while predictive analytics anticipates the future. Business intelligence involves using data to support decision making and aims to provide historical, current and predictive views of business. As technologies advanced, business intelligence evolved from being organized under IT to potentially being aligned under strategy management. Effective communication between business and analytics professionals is important for organizations to benefit from predictive analytics. The business case for predictive analytics includes enabling strategic planning, competitive analysis, and improving business processes to work smarter.
Business analytics workshop presentation finalBrian Beveridge
This document outlines an agenda and presentation for a business analytics seminar for credit union executives and board directors. The presentation will define business analytics, explain how it can help credit unions address key issues like margin compression and regulatory compliance, and provide examples of how analytics can be applied to areas like marketing, risk management, and branch performance. Attendees will learn how predictive analytics can help credit unions retain members, optimize pricing, and streamline operations. The presentation will also cover getting started with business analytics projects.
The document provides an overview of business analytics (BA) including its history, types, examples, challenges, and relationship to data mining. BA involves exploring past business performance data to gain insights and guide planning. It can focus on specific business segments. Types of BA include descriptive analytics like reporting, affinity grouping, and clustering, as well as predictive analytics. Challenges to BA include acquiring high quality data and rapidly processing large volumes of data. Data mining is an important task within BA that helps handle large datasets and specific problems.
Credit card fraud detection using python machine learningSandeep Garg
This document provides an overview of machine learning tools, technologies, and the data preparation process. It discusses collecting and selecting relevant data, data visualization, labeling data for supervised learning, and transforming raw data into a tidy format. The document also covers various data preprocessing techniques, including data cleaning, formatting, handling missing values and outliers, smoothing, aggregation, generalization, and data reduction methods. The goal of these preprocessing steps is to prepare raw data into a structured format suitable for machine learning modeling.
The Business Analytics Value PropositionEric Stephens
Presentation made to the Nashville Technology Council Analytics Peer Network meeting on May 30, 2013. Discussion of the impact of analytics to an organization, along with use cases that can help convey the value of the practice to executives and other managers.
The speaker discusses the importance of evaluating big data analysis to improve projects. They recommend getting a second opinion on methodology or using another data source for verification. A case study on estimating traffic congestion is presented where ground truth sensor data was collected and compared to estimates using metrics to provide feedback. Lessons include ground truth not being easy to obtain, using the right tools like Python and pandas for evaluation, and having an iterative workflow for timely feedback.
System Dynamics, Analytics & Big Data (16th Conference of the UK Chapter of t...Michael Mortenson
The document discusses the relationship between analytics, big data, and system dynamics. It begins by providing background on the growth of analytics and big data. It then discusses relationship problems between analytics and operations research. The main part of the document introduces the Dianoetic Management Paradigm to describe the evolution of management thinking and related technologies over time. It also describes categories of analytics from descriptive to predictive to prescriptive. Finally, it discusses implications of big data for system dynamics, including opportunities around high-volume, unstructured, and streaming data as well as related technologies.
This document provides tips for aspiring data scientists. It advises them to start by focusing on a topic that interests them and to clearly define their objectives and data collection process. It also recommends that they visualize their data, understand the context, look for additional insights, evaluate results, and find effective uses of the data. The document notes that data is becoming increasingly important in all industries and companies without data-savvy managers will be at a disadvantage.
The document discusses business analytics and decision making. It defines key concepts like data warehousing, data mining, business intelligence, descriptive analytics, predictive analytics, and prescriptive analytics. It explains how these concepts are used to extract insights from data to support decision making in organizations. Examples of how different types of analytics can be applied in a retail context are provided.
This presentation introduces big data and explains how to generate actionable insights using analytics techniques. The deck explains general steps involved in a typical analytics project and provides a brief overview of the most commonly used predictive analytics methods and their business applications.
Vijay Adamapure is a Data Science Enthusiast with extensive experience in the field of data mining, predictive modeling and machine learning. He has worked on numerous analytics projects ranging from healthcare, business analytics, renewable energy to IoT.
Vijay presented these slides during the Internet of Everything Meetup event 'Predictive Analytics - An Overview' that took place on Jan. 9, 2015 in Mumbai. To join the Meetup group, register here: http://bit.ly/1A7T0A1
[Ai in finance] AI in regulatory compliance, risk management, and auditingNatalino Busa
AI to Improve Regulatory Compliance, Governance & Auditing. How AI identifies and prevents risks, above and beyond traditional methods. Techniques and analytics that protect customers and firms from cyber-attacks and fraud. Using AI to quickly and efficiently provide evidence for auditing requests.
Data analytics refers to the broad field of using data and tools to make business decisions, while data analysis is a subset that refers to specific actions within the analytics process. Data analysis involves collecting, manipulating, and examining past data to gain insights, while data analytics takes the analyzed data and works with it in a meaningful way to inform business decisions and identify new opportunities. Both are important, with data analysis providing understanding of what happened in the past and data analytics enabling predictions about what will happen in the future.
Material for the 26 Oct 2015 lecture I held for Aalto University business students. The lecture focuses on the high level topics in analytics and Big Data that are either central to the subject or just highly visible in the media.
The main messages of the lecture are:
- The purpose of analytics and of the data analyst is to solve business problems
- Big Data brings over some very special traits to doing analytics that don't exist when working working with smaller datasets. Understanding these traits is a must for successful analytics.
- Deploying analytics is more dependent on humans than on technology
- Data and analytics are nowadays significant assets to many companies. Therefore they need their own strategy and need to be managed just like any other business critical assets.
Learn about the emerging field of big data and advanced quantitative models and how the Rady School's MS in Business Analytics program is designed to solve important business problems.
HYDROLIFT 24, 2003, £49,995 For Sale Brochure. Presented By yachtingelite.comWolfgang Stolle
HYDROLIFT 24
Puerto Banus, Spain
Hyrdolift 24 With Single Mercruiser Inboard 496 Ho Petrol Engines (430hp). This Particular Model Is In Fantastic Condition, Benefitting From A New Paint Job, New Engines (2 Hours Use) And New Outdrives. In Fantastic Condition And Offers Would Be Seriously Considered. Please Contact A Member Of Our Team For Further Information Or Indeed To Arrange A Viewing.
For Sale Brochure. Presented By yachtingelite.com. Visit Site http://www.yachtingelite.com by yachtingelite.com. Real Estate,Luxury Property
как удержать В2В клиентов с помощью изучения целевой аудиторииEkaterina Gould
основываясь на опыте 14ти лет работы в сфере В2В, мы сформулировали эти принципы работы с клиентами. Эмоциональная связь с потребителем, регулярная работа с CRM базой, знание поведенческих характеристик клиента. В нынешних экономических условиях Client is a King! дайте то, что нужно только ему.
Our approach to enterprise architecture at Kiwibank 2008 presented to Microsoft Architects Council sessions in Melbourne, Canberra, Sydney, Wellington, Christchurch, and Auckland.
Ravi Hariani's professor wrote a letter of recommendation for his pursuit of a career in data analytics. The professor taught Ravi in two graduate MIS courses, where Ravi demonstrated a broad understanding of data science and in-depth knowledge of data mining algorithms. Ravi completed projects using tools like SQL Server, Excel, SPSS, R, Python and Tableau. The professor found Ravi to be open-minded, willing to take a lead role, and eager to explore new ideas and learn. The professor described Ravi as well-balanced, a great team player with strong communication skills and results-oriented.
Kiwibank has embraced Enterprise 2.0 by implementing collaboration tools and social computing to address increasing infrastructure complexity. As staff and services grew from 2001-2009, Kiwibank deployed unified communications, cloud hosting, and subscriptions. The bank's experience shows that a social approach focusing on individuals, transparency through public metrics, and over 80% user-generated content answer rates has supported growth. Business intelligence is also decentralized through data modeling and analytics, with system monitoring and modeling moving from development to production use.
SystemT: Declarative Information Extraction (invited talk at MIT CSAIL)Laura Chiticariu
Invited talk at MIT CSAIL, March 8 2016
Information extraction (IE), the task of extracting structured information from unstructured or semi-structured data, is increasingly important to a wide array of enterprise applications, ranging from Business Intelligence to Data-as-a-Service. Such applications drive the following main requirements for IE systems: accuracy, scalability, expressivity, transparency, and customizability.
SystemT, a declarative IE system, has been designed and developed to address these requirements. It is based on the basic principle underlying relational database technology: complete separation of specification from execution. SystemT uses a declarative language for expressing NLP algorithms called AQL, and an optimizer that generates high-performance algebraic execution plans for AQL rules. It makes IE orders of magnitude more scalable and easy to use, maintain and customize.
SystemT ships today with multiple products across 4 IBM Software Brands. Furthermore, SystemT is used in multiple ongoing research projects and being taught in universities. Our ongoing research and development efforts focus on making SystemT more usable for both technical and business users, and continuing enhancing its core functionalities based on natural language processing, machine learning, and database technology.
Time Difference: How Tomorrow's Companies Will Outpace Today'sInside Analysis
The Briefing Room with Mark Madsen and WebAction
Live Webcast Feb. 10, 2015
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=fa83c6283de99dfb6f38b9d7199cb452
In our increasingly interconnected world, the windows of opportunity for meaningful action are shrinking. Where hours once sufficed, minutes are now the norm. For some transactions, seconds make all the difference, even sub-seconds. Meeting these demands requires a new approach to information architecture, one that embraces the many innovations that are fundamentally changing the data-driven economy.
Register for this episode of The Briefing Room to hear veteran Analyst Mark Madsen of Third Nature as he explains how a confluence of advances are changing the nature of data management. He'll be briefed by Sami Akbay of WebAction, who will showcase his company's real-time data platform, designed from the ground up to meet the challenges of leveraging Big Data in concert with all manner of operational enterprise systems.
Visit InsideAnalysis.com for more information.
1) In-memory computing is growing rapidly, with the total data market expected to grow from $69 billion in 2015 to $132 billion in 2020.
2) In-memory databases are gaining popularity for applications that require fast response times, like telecommunications and mobile advertising, as memory access is faster than disk access.
3) Modern applications are driving adoption of in-memory solutions as they generate more data from more users and transactions and require faster performance to handle growing traffic.
4) Two examples presented were DellEMC using MemSQL for a real-time customer 360 application and an IoT logistics application called MemEx that processes sensor data from warehouses for predictive analytics.
Splunk provides an operational intelligence platform that allows organizations to:
1. Ingest data from various sources in real-time for searching, visualizing, and analyzing.
2. Establish a culture of continuous improvement through continuous delivery, continuous insights, and fast feedback to move at market speed.
3. Build a strong IT foundation with hybrid cloud, continuous delivery of applications, and continuous insights from proactive monitoring to enable business agility.
A Journey Through The Far Side Of Data Sciencetlcj97
This document summarizes a presentation on data science and artificial intelligence. It discusses how AI is transforming businesses in many ways, including automating repetitive tasks, improving customer experiences, and driving revenue growth. It also mentions that while data is important, AI is needed to transform organizations through intelligent process optimization and innovation. The document provides examples of how various companies are applying AI in sales, customer service, and other areas. It emphasizes that AI strategies should focus on innovation, identifying high-impact use cases, and developing people's data science skills.
Businesses that take data seriously organise themselves around data, treating it as a valuable organisational asset.
The emerging trends in digital analytics and the decision points companies face when shifting from siloed departmental analytics to company-wide shared insights.
This talk addresses hitting the limits of what businesses can do in batch data processing and common patterns that accelerate their decisioning using real time.
Enterprise Grade Data Labeling - Design Your Ground Truth to Scale in Produ...Jai Natarajan
We describe why and how to be mindful about designing you data annotation pipeline to be scalable and to delivery consistent high quality results regardless of domain
Emerging Prevalence of Data Streaming in Analytics and it's Business Signific...Amazon Web Services
Learning Objectives:
- Get an overview of streaming data and it's application in analytics and big data.
- Understand the factors driving the accelerating transformation of batch processing to real-time.
- Learn how you should plan for incorporating data streaming in your analytics and processing workloads.
Business can now easily perform real-time analytics on data that has been traditionally analyzed using batch processing in data warehouses or using Hadoop frameworks, and react to new information in minutes or seconds instead of hours or days. In this webinar, Forrester analyst Mike Gualtieri and Amazon Kinesis GM Roger Barga will discuss this prevalent trend, it's business significance, and how you should plan for it. You will also learn about the AWS services that can help you get started quickly with real-time, streaming applications fore your analytics and big data workloads.
Moving Targets: Harnessing Real-time Value from Data in Motion Inside Analysis
The Briefing Room with David Loshin and Datawatch
Live Webcast Feb. 17, 2015
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=4a053043c45cf0c2f6453dfb8577c72a
Patience may be a virtue, but when it comes to streaming analytics, waiting is no option. Between Big Data and the Internet of Things, businesses are faced with more data and greater complexity than ever before. Traditional information architectures simply cannot support the kind of processing necessary to make use of this fast-moving resource. The modern context requires a shorter path to analytics, one that narrows the gap between governance and discovery
Register for this episode of The Briefing Room to hear veteran Analyst David Loshin as he explains how the prevalence of streaming data is changing business pace and processes. He’ll be briefed by Dan Potter of Datawatch, who will tout his company’s real-time data discovery platform for data in motion. He will show how self-service data preparation can lead to faster insights, ultimately fostering the ability to make precise decisions at the right time.
Visit InsideAnalysis.com for more information.
Functionalities in AI Applications and Use Cases (OECD)AnandSRao1962
This presentation was given at the OECD Network of AI Specialists (ONE) held in Paris on February 26 and 27. It covers the methodology for assessing AI use cases by technology, value chain, use, business impact, business value, and effort required.
Fluturas presentation @ Big Data Conclavefluturads
This document discusses 5 case studies of companies using big data to solve real-world problems:
1. A telecom company used machine data from perimeter devices to detect security patterns and reduce network threats from hackers and internal activists.
2. An online travel company used customer behavior data to understand travelers' intent and improve customer experience and cross-selling.
3. A company used telecom logs to detect signals in watched lists to enhance national security and prevent threats.
4. Mobile data was analyzed to spot friction points in a travel company's mobile funnel and drive more revenue.
5. Money transmission patterns were modeled using a graph database to minimize fund leakages to watched entities.
Bridging the Gap: Analyzing Data in and Below the CloudInside Analysis
The Briefing Room with Dean Abbott and Tableau Software
Live Webcast July 23, 2013
http://www.insideanalysis.com
Today’s desire for analytics extends well beyond the traditional domain of Business Intelligence. That’s partly because business users are realizing the value of mixing and matching all kinds of data, from all kinds of sources. One emerging market driver is Cloud-based data, and the desire companies have to analyze this data cohesively with their on-premise data sets.
Register for this episode of The Briefing Room to learn from Analyst Dean Abbott, who will explain how the ability to access data in the cloud can play a critical role for generating business value from analytics. He’ll be briefed by Ellie Fields of Tableau Software who will tout Tableau’s latest release, which includes native connectors to cloud-based applications like Salesforce.com, Amazon Redshift, Google Analytics and BigQuery. She’ll also demonstrate how Tableau can combine cloud data with other data sources, including spreadsheets, databases, cubes and even Big Data.
EVOLVING PATTERNS IN BIG DATA - NEIL AVERYBig Data Week
The document discusses evolving patterns in big data usage, including enterprise data caching using massive key-value stores, enterprise messaging pipes using Kafka, and NoSQL as a service. It also covers data lakes for centralized raw data storage and Lambda architecture for near real-time and batch processing. Current trends include growing Cassandra usage, Kafka for scalable messaging, and containerization and cloud adoption. Future areas may include graph databases, Spark evolution, and data virtualization.
Enabling digital business with governed data lakeKaran Sachdeva
Digital business is enabled by Artificial intelligence, Machine learning, and data science. Artificial intelligence and machine learning are dependent on right Information architecture and data foundation. Governed data lake infused with governance and data science platform gives you the power to take the organization in the digital transformation and AI journey.
TechWise with Eric Kavanagh, Dr. Robin Bloor and Dr. Kirk Borne
Live Webcast on July 23, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=59d50a520542ee7ed00a0c38e8319b54
Analytical applications are everywhere these days, and for good reason. Organizations large and small are using analytics to better understand any aspect of their business: customers, processes, behaviors, even competitors. There are several critical success factors for using analytics effectively: 1) know which kind of apps make sense for your company; 2) figure out which data sets you can use, both internal and external; 3) determine optimal roles and responsibilities for your team; 4) identify where you need help, either by hiring new employees or using consultants 5) manage your program effectively over time.
Register for this episode of TechWise to learn from two of the most experienced analysts in the business: Dr. Robin Bloor, Chief Analyst of The Bloor Group, and Dr. Kirk Borne, Data Scientist, George Mason University. Each will provide their perspective on how companies can address each of the key success factors in building, refining and using analytics to improve their business. There will then be an extensive Q&A session in which attendees can ask detailed questions of our experts and get answers in real time. Registrants will also receive a consolidated deck of slides, not just from the main presenters, but also from a variety of software vendors who provide targeted solutions.
Visit InsideAnlaysis.com for more information.
ADV Slides: What the Aspiring or New Data Scientist Needs to Know About the E...DATAVERSITY
Many data scientists are well grounded in creating accomplishment in the enterprise, but many come from outside – from academia, from PhD programs and research. They have the necessary technical skills, but it doesn’t count until their product gets to production and in use. The speaker recently helped a struggling data scientist understand his organization and how to create success in it. That turned into this presentation, because many new data scientists struggle with the complexities of an enterprise.
1. Visa operates the world's largest payment network, VisaNet, which processes over 130 million authorizations per day across 28 million merchant locations.
2. Hadoop offers the potential to analyze Visa's large volumes of transaction data faster than currently possible. Visa is researching how to configure Hadoop solutions and integrate them securely within its existing financial systems.
3. Key questions include how to securely load data into Hadoop clusters, integrate security features like hardware security modules, and ensure solutions meet Visa's requirements of being fast, secure, flexible, reliable and scalable for enterprise use.
Agile Data Science is a lean methodology that is adopted from Agile Software Development. At the core it centers around people, interactions, and building minimally viable products to ship fast and often to solicit customer feedback. In this presentation, I describe how this work was done in the past with examples. Get started today with our help by visiting http://www.alpinenow.com
How to Use Artificial Intelligence by Microsoft Product ManagerProduct School
The talk focused on the Fundamentals of Product Management, leveraging the speaker's personal experiences in the AI field. It covered core Product Manager topics such as managing customer needs, business goals & technology feasibility, the holy trinity of the Product Manager discipline, delve into data analyses, rapid experimentation, and execution, and finally, explored the challenges of customer privacy, bias, and inclusivity in AI products.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
3. My Technology Office Challenges
Product /
Marketing /
Operations
Technology
Delivery
Architecture Narrative
1
– what, why…
2
Make the Right Decisions
3
Capability Development
4
Derive Value from Data
4. 2. Making the right decision
Forces for control
• Processes and tools
• Comprehensive documentation
• Plans
Forces for creation
• Individuals and Interactions
• What works beats theory
• Change is everywhere
6. My observations
Our best work has happened when 3 things were true:
1. We give it a go
2. With a supportive business team that’s not scared to
understand the little details
3. And a committed and passionate technical lead
4. That ensures people work together.
8. Architecture Briefing Forum
•
•
•
•
•
•
•
•
•
•
•
Distributed configuration
across Internet Banking
servers
Federated Identity
Anti-Money Laundering
Operations Management
with System Center
Web Service Gateways
Application Build and
Deployment
Storage
Pre-Production Update
NoSQL logging design
comparing MongoDB and
RavenDB
Crypto 101
…
9. 4.Gaining value from information
Key insight has been that
multiple groups are
required to gain full
benefit.
Why?
Analytics platform
Enterprise
Information platform
Differing skills and
knowledge.
Data Platform
IT has
a role
across
all
10. Our Data Platforms
Storage
Database
Hitachi Virtualised Storage
½ PB of data
= 7 years of HD-TV
1/5th is database storage
High growth rate
• Microsoft SQL Server
• NoSQL: Universe, RavenDB, MongoDB
800
600
Shared Storage (TB)
Staging
• 65GB exposed to systems every day
• >1000 data jobs / day
400
200
0
2000
2005
2010
2015
12. Our Enterprise Information Platform
• Key strategy is to unlock information assets
• Make data available into Analyst Playpens
• Through improved data acquisition / transformation /
delivery
• Currently in progress…
13. Our Analytics Platform
Microsoft Excel (+ VBA/.NET)
Tibco S+
Microsoft SQL Server (DMX)
Data
Sungard ALM
Analyst’s
Playpen(s)
All different.
All with strengths.
All suited to different users.
14. And Analytics Means What Exactly?
1. Optimisation
Optimal decisions under competing constraints.
2. Understanding connectivity
Connectivity between attributes leading to a decision
ie “market basket analysis”; or understanding clustering
of outcomes.
3. Predicting the future
Predicting future outcomes based upon history.
25. Online buyer behaviour
Is online purchasing more or less important in the regions
versus the cities?
My hypothesis is yes
High
Low
Can’t disprove it…
29. What do we see?
Large variances across companies…
Minor regional differences…
Colours = Companies
$0
$50
$100
Key data is power company avg bill…
Optimisation
possibility of about
$15/month or
$180/year
$80
$90
$100 $110 $120 $130 $140 $150 $160 $170
30. Metrics and Analytics for IT
• Summer of Tech project
• To build digital signage that displays
metrics across operations, delivery,
and the business.
http://www.summeroftech.co.nz/
• Focused on the IT consumer
31. Summary
We build narratives to explain what and why.
We work with all of IT to grow capability and
ensure we’re collectively making the right
decisions.
We support analysts by ensuring they have the
platforms, tools, and skills to succeed.
Editor's Notes
There are 4 technology leadership challenges I face at kiwibank of which the most important is the Architecture narrative.The single most important one is creating understanding through narrative. It’s about explaining what we have, why we have and what we’ve got to do. Without this we’d end up with disjoint, disconnected strategies and you see this time and again in companies and it’s easy to identify.Replicated systemsDiversity of technology without good reasonA collection of ‘point in time’ solutions.I don’t want this to happen at Kiwibank; it’s critical we keep moving ahead with the context of what and why flowing through everything we do.The second challenge I face is meeting the knowledge gap.My GM of IT is constantly remarking on the increasing complexity of the technologies we deal with. And I have to ask myself is it true? Is IT more complex now than it was in the 90s?I remember the 90s quite clearly…We wrote web applications in .asp vbscript/jscriptWe used local disk a lot and didn’t worry about SANs or storage virtualisationWe racked servers that looked like PCs and they were largely self containedToday it is different.Infrastructure is all about virtualization, all about adding another layer of abstraction to move away from discrete computing boxes and move to private cloudsOperating systems change and we have very good examples affecting us all today in the client and server space with Windows, iOS and Android.Software is about buses, events, asynchronous messaging and distribution across cores, CPUs and parallel computingDevelopment has become more refined. Concepts that were only talked about in math and computer science departments 20 years ago are now the stuff of every day practice.Functional programming is in along with concepts like Monads, combinators and higher order functionsThe role of the relational database is now better understood so we have regained appreciation of the many other types of databases out there. Each of these has a niche within which they makes sense and it’s the reason for the nosql movement.So keeping up with recent developments is a big task and we can’t avoid it or fear it – being good at learning and applying these technologies is key to being competitive.Unfortunately, the old model of training courses doesn’t really seem so useful to me. The content is not what I think we need. So for me it’s about building this knowledge within the bank. It’s about creating networks that bring new technologies in and finding ways to foster the spread within our IT function.
We’re lucky at Kiwibank, nearly all the technologists are located on one floor. We’re not split across multiple buildings. The most important thing for us is to work together.And there’s two fundamental approaches: eitherpush down control through processes, tools and enforced planning and documentation, or work with the individuals creating the future to ensure it’s right.This language shouldn’t seem strange. It should sound familiar.
Because it’s straight out of the Agile Manifesto.With our co-location we’re set up for an agile approach. We don’t have staff spread across different countries, different timezones, different languages. We’re all within walking distance of each other so it’s important we capitalize on it.The key technique that Agile uses in software to make sure you’re on track is to prefer working software over theory. In the software development world this is all about unit testing and iron out bugs early with short release cycles.When it comes to technology governance you can take the same approach.
Why are we doing xyz? Recognise the missing gap of the context – which is what we need to strengthenThese are my observations.We’re we’ve done best is when 4 things have happened.We’ve given something a go quickly – not over a year, not in a big programme, but instead in a brief experiment.With the backing of a business team that aren’t scared to get into the nitty gritty detail (ie not focused on vendor packages)In association with a committed, capable technical leadThat ensures we all work together.We have examples: Our approach to AIRB Our approach to refactoring InTouchWe’ve recently done some analysis on a system that we use within our sales & service function. It consists of a lot of code and it’s been running for a long time. This is an example of a lesson learnt – it’s not all rosy. We deployed quickly and got early benefits but over the years we’ve transitioned the lead development role across many people and what we’ve found is that context for technical decisions had been lost, business driven change had then resulted in lots of poor implementation decisions and the code had degraded.To address it we’re doing two things. We’re spending some money on refactoring but the key thing for me is that we’ve really identified that the lead developer needs our support. She has to have the mandate to make deeply technical decisions and to ensure that we do it right in the future. I don’t want to have to transition to another technical lead and undergo the same discovery process all over again.And respect the importance of individualsAny examples?
Next problem is capability development.Wehave 3 technical forums at Kiwibank.The Architecture Briefing Forum is the longest running and most attended. The Enterprise Information Forum is still clarifying it’s role but it’s become apparent its main focus is on data management. The Analytics Forum is new. It’s become clear that the topics the analysts want to discuss aren’t really applicable to the others and there’s enough content and interest to make something of it in isolation of the others.As a case study the ABF is fascinating. It’s on Friday mornings – so one of the most relaxed times of the week. Topics are varied – let’s look at a few recently.
Participation occurs across all groups, but realistically there is a much higher participation from the infrastructure delivery team.Personally I like the developer focussed material but I have to admit it doesn’t tend to keep the audience so I can see another spin off in the developer space happening as we grow.The clear message is that the forum provides a fantastic opportunity to socialise ideas, technologies, new ways of working.
Which brings us to the last challenge.Delivering value from data.We’ve all got large business intelligence teams right?We’ve all got large sources of corporate data?And you’ve heard of big data, and business analytics, but you’re wondering what it’s about and how to get there?Well me too.I’m fascinated by data. My background in science was a while ago but I think this part of it must be in my psyche. I feel happy manipulating information to test ideas or discover relationships.What’s clear to me is that deriving value from analytics requires investment in our underlying data infrastructure.Data PlatformInfrastructure heavy skill baseHigh cost to manage over timeKey challenges today are distributed data management, de-duplication, backup, archiving, IO, tieringEnterprise Information PlatformThis is a big problem area. Classic BI/MIS/EI – it’s all about data extraction, transformation, aggregation, management into playpens and reporting. Big problems today are volumes, lineage, impact analysis, master data and metadata management.Analytics PlatformThis is about catering to analysts – don’t get in the way, instead help them to do good work. Recognising that algorithms are diverse and implementation technologies suited to those algorithms are also diverse. Classic tool sets like SAS are now supplemented by tools like R and Hadoop. In this realm code is ever present – you can’t do analytics if you’re not going to accept having and doing code.The question is… what’s IT’s role in this environment?
Unstructured business data.Clearly IT has a major role in storage.Virtualised storage layer that underlies our server infrastructure.Critical component of our Private Cloud infrastructure as it underlies our server virtualisation environment.It’s a lot of shared storage.http://mozy.com/blog/misc/how-much-is-a-petabyte/1PB is 13.3 years of HD-TV
What’s the story?What’s all this exotic technology about.Everyone will presumably be familiar with the big database brands like DB2, Oracle and SQL Server but have you heard about these?The thing is this stuff is new and it’s what’s driving the internet scale application.Facebook, Pinterest, Instagram, Twitter – these apps don’t use big brand databases.Instagram – EC2Twitter – FlockDB (distributed fault tolerant graph database)Facebook – Memcache (9000 instances), MySQL atomic storage (4000 shards) according to http://gigaom.com/cloud/facebook-trapped-in-mysql-fate-worse-than-death/Pinterest – EC2So, OK these are obviously useful technologies when operating at internet scale. But do they have a place within a bank in the NZ retail market?Actually, yes. There are plenty of sensible deployment scenarios where these technologies can supplement our predominantly SQL Server database environment with features that address edge cases.Here’s one: distributed configuration management for which we use RavenDB.Here’s another: message auditing within our service infrastructure.And yet one more we’re investigating for the future: a customer communications document store. The current solution based on a relational database has worked well up till now but the immense size of our customer comms now, and obvious size of the infrastructure means we’re now looking for a new approach. Document stores are one of the most attractive use cases for nosql – you have the potential to get a high performing, resiliant database built on commodity hardware.
And clearly supporting enterprise information and all required databases and tools.
But with analytics it’s not quite so obvious. The analytics platform is characterised by a greater diversity of technology and tooling than any other functional area in IT.We have S+ for model development.We have SAS for model implementation and reporting.We have Excel and VBA.We have .NET apps.We have R, the opensource version of S+, favoured tool of the NY Times infographics department and 500 google statisticians (http://blog.revolutionanalytics.com/2012/07/another-r-mention-in-the-nyt.html, http://blog.revolutionanalytics.com/2012/07/applications-of-r-at-google.html)We have SQL Server including the data mining extensions embedded in SQL.We have Predixion Insight in use by the marketing analysts which provides an easy to use interface to SQL Server data mining via Excel.This is complicated stuff!And it get’s much more complex as data volumes increase and we have to split calculations across many machines. The classic example of this being Hadoop.It’s not easy to reduce that diversity. It’s there because invariably you can’t get everything you need from one tool. There are fundamental reasons for this. Data storage is a big one. Excel, R, S+, any RDBMS – they’re all optimised by design for smaller data sets. For larger data you typically need a different approach. SAS gets you part way and then you’re into the world of big data for which vendors are now producing packaged tooling.
OptimisationSometimes called prescriptive – defining the best course of action for the future; always associated with a need to maximise or minimise an outcome; often time or profit.Classic optimisation problems are:Portfolio management to maximise returnScheduling to minimise costUnderstandingThe discovery of connections, the uncovering of relationships. What attributes influence an outcome. Attributes that often associate together (aka market basket analysis). Clusters of results.PredictionPredicting future outcomes based upon history. When x and y equal these values then the result is always this value.
A lot of technologies. A lot to learn. These systems all have domains within which they excel so you can’t just rely on one tool.
Most of the analytics software is pretty much free.But when you start processing larger quantities of data, especially when the number of potential variables goes up then resource requirements can grow enormously. The great promise of big data is associated with the potential for great cost and great infrastructure.Let’s make it tangible: text mining email communication. My email amounted to 1.1GB last year (2011). Assuming 10kb/msg that’s roughly 115,000 messages. To start mining this I need to construct a matrix with 115,000 columns and the row count will be whatever number of distinct terms there are – commonly over a 100,000 for a corpus of documents this size. This is too big for a workstation. A fairly powerful workstation can process maybe 10,000 short documents in a practical amount of time. So you need to split the task up across many machines.Hadoop is the archetypal example: an opensource framework for splitting aggregation type tasks across many machines but it’s complex to implement and it’s not easy to use.Vendors like IBM, SAS (and soon Microsoft) are making this much easier but there’s nothing out there that makes these tasks magic. You need math grads to make sense of it.And for the infrastructure you really have to turn to the cloud.
Hmmm, not sure when I changed that setting but I soon discovered it.
This is the result of me running through a text analysis problem on a cloud service from Microsoft called Cloud Numerics. It was for a text analysis problem.What I did wrong here was I left it running. User error – decommission the cloud infrastructure you assign or else you’ll get charged.In fact, in this case if I’d de-provisioned the infrastructure once the job had come to an end the cost would’ve been $0.So firstly, be careful using cloud infrastructure but secondly, if you use it right it makes a lot of financial sense for analytics workloads.And actually, for these problems it’s very secure – if you do it right, most of these problems can be set up such that you’re not exposing any potentially sensitive data and the data governance issues over cloud ownership can be avoided.
A few weeks ago during a vendor presentation one of the business team mentioned sentiment while we talking about that vendor’s data products. Sentiment’s a hot topic on the internet at the moment, if you search you’ll find many examples and the approach goes basically like this: score your text based upon a list of known words or phrases with previously derived sentiment scores.To get the previously derived sentiment scores you have a few options. You’ll find freely available word lists online, or you can construct your own. You can score the terms from a –ve number to +ve number or you can just create two lists: one of negative words; and one of positive words. One way to get this yourself in more automated fashion is to leverage customer product reviews, again these are also available freely online from a variety of sources. Keep in mind that regional phrases/terms will have a bearing on the analysis so in our case I accounted for a few common NZ terms.Applying my list of positive and negative sentiment terms against Kiwibank’s online relationship manager messages gave me a reasonable response straight away and we could clearly correlate low points and high points. I can’t show that here but I can show an equivalent exercise performed for public tweets off Twitter that reference @ASBBank, @BNZBank and @KiwibankNZ.
A few weeks ago during a vendor presentation one of the business team mentioned sentiment while we talking about that vendor’s data products. Sentiment’s a hot topic on the internet at the moment, if you search you’ll find many examples and the approach goes basically like this: score your text based upon a list of known words or phrases with previously derived sentiment scores.To get the previously derived sentiment scores you have a few options. You’ll find freely available word lists online, or you can construct your own. You can score the terms from a –ve number to +ve number or you can just create two lists: one of negative words; and one of positive words. One way to get this yourself in more automated fashion is to leverage customer product reviews, again these are also available freely online from a variety of sources. Keep in mind that regional phrases/terms will have a bearing on the analysis so in our case I accounted for a few common NZ terms.Applying my list of positive and negative sentiment terms against Kiwibank’s online relationship manager messages gave me a reasonable response straight away and we could clearly correlate low points and high points. I can’t show that here but I can show an equivalent exercise performed for public tweets off Twitter that reference @ASBBank, @BNZBank and @KiwibankNZ.
Text mining is the common term for the a process of uncovering word relationships to create an understanding of topics. The more technical term would be semantic analysis. The even more technical one is Latent Semantic Analysis.The basic idea is to count the frequency of terms in documents and understand what terms associate together. Hopefully the associated terms will correlate to an obvious topic area and you could imagine a banking topic to be maybe ‘customer analysis of home loan deals’, or perhaps ‘a customer has a problem’ and so on…This is a hot topic itself so when you start searching on the net you’ll find a number of articles. Again R examples abound – particularly from the R text mining package ‘tm’ but also from products like SAS which has a text mining module.As a process the steps are roughly: Collect the textRemove stopwords – the words of no valueDeal with any synonymsTransform related word forms to a common format egrunning and run mean much the sameRun the algorithms that decompose the topics.The algorithm side is both complex (singular value decomposition) and a numerically intensive task. This is classic big data territory where splitting the work across machines becomes useful. However, we can make slow progress on a single workstation if we keep to a few thousand small documents. Let’s use Twitter again.Let’s now look at Tweet topic
Note that on my screen these connectivity plots are interactive so I can drag words around and re-organize as seems sensible.Next steps could be … Classify/predict eg emotional state: happy | sad | angry
Maybe look at online purchasing usage in regions versus cities? We can see these transactions by virtue of attributes like reference numbers, card merchant names etc.Let’s normalize for regional population.What do we find? Online buying is much more important for rural regions especially the south island’s west coast and southland. As for Malborough… I have no idea.
A random sampling of our transactions from a day earlier this year colour coded to indicate the value of the payment and plotted across great circle arcs. This is a clever little trick that comes from the R community and was famously applied in the facebook IPO.
And if you don’t want to put the effort into figuring this out yourself – you can get others to do it for you.What you see here is the homepage for kaggle – a site that supports competition driven analytics. You can see an example at the bottom of the page; $100,000 up for grabs to develop a scoring algorithm for student examination with 151 teams competing for the prize.
Strongly advise going to whatsmynumber.org.nz and following the Electricity Authorities advise.They predict average annual benefits of $165 but it seems to me that many of our customers could gain $15/month or $180 annual benefit.