Hear how Walmart has used Elasticsearch to ingest more than 4 billion metadata records and use security, alerting, and Graph features to identify fraudulent gift-card requests in real time.
Henry Peyret Presentation - Data Governance 2.0.
Based on the analysis of Digital Transformation and Values Transformation, Forrester gives its insight and orientations in terms of Data Governance 2.0 and Data Citizenship.
data management, information management, data, big data, personal organization, organization, file management, scientific research, research, project management, data security, file naming conventions, data management plan,
This describes a conceptual model approach to designing an enterprise data fabric. This is the set of hardware and software infrastructure, tools and facilities to implement, administer, manage and operate data operations across the entire span of the data within the enterprise across all data activities including data acquisition, transformation, storage, distribution, integration, replication, availability, security, protection, disaster recovery, presentation, analytics, preservation, retention, backup, retrieval, archival, recall, deletion, monitoring, capacity planning across all data storage platforms enabling use by applications to meet the data needs of the enterprise.
The conceptual data fabric model represents a rich picture of the enterprise’s data context. It embodies an idealised and target data view.
Designing a data fabric enables the enterprise respond to and take advantage of key related data trends:
• Internal and External Digital Expectations
• Cloud Offerings and Services
• Data Regulations
• Analytics Capabilities
It enables the IT function demonstrate positive data leadership. It shows the IT function is able and willing to respond to business data needs. It allows the enterprise to meet data challenges
• More and more data of many different types
• Increasingly distributed platform landscape
• Compliance and regulation
• Newer data technologies
• Shadow IT where the IT function cannot deliver IT change and new data facilities quickly
It is concerned with the design an open and flexible data fabric that improves the responsiveness of the IT function and reduces shadow IT.
Why apache Flink is the 4G of Big Data Analytics FrameworksSlim Baltagi
Apache Flink is a community-driven open source and memory-centric Big Data analytics framework. It provides the only hybrid (Real-Time Streaming + Batch) open source distributed data processing engine supporting many use cases.
Flink uses a mixture of Scala and Java internally, has very good Scala APIs and some of its libraries are basically pure Scala (FlinkML and Table).
At its core, it is a streaming dataflow execution engine and it also provides several APIs for batch processing (DataSet API), real-time streaming (DataStream API) and relational queries (Table API) and also domain-specific libraries for machine learning (FlinkML) and graph processing (Gelly).
In this talk, you will learn in more details about:
What is Apache Flink, how it fits into the Big Data ecosystem and why it is the 4G (4th Generation) of Big Data Analytics frameworks?
How Apache Flink integrates with Apache Hadoop and other open source tools for data input and output as well as deployment?
Why Apache Flink is an alternative to Apache Hadoop MapReduce, Apache Storm and Apache Spark? What are the benchmarking results between Apache Flink and those other Big Data analytics frameworks?
Overview of Data Loss Prevention Policies in Office 365Dock 365
Presentation about identifying, monitoring, and automatically protect sensitive information across Office 365.
With a DLP Policy, you can:
- Identify sensitive information across many locations, such as SharePoint Online and OneDrive for Business.
- Prevent the accidental sharing of sensitive information.
- Monitor and protect sensitive information in the desktop versions of Excel 2016, PowerPoint 2016, and Word 2016.
- Help users learn how to stay compliant without interrupting their workflow.
- View DLP reports showing content that matches your organization's DLP policies.
Visit www.mydock365.com to learn more about SharePoint with Dock.
Migrating 3000 users and 1100 applications from Lotus Notes to Office 365Arno Flapper
A large international customer decided to move into Microsoft Office 365. Their challenge is to migrate 3000+ users, 1100 applications and 100 Tb of file share data located in 50 locations and datacenters across Europe. With this migration all datacenters must be replaced by a central datacenter in The Netherlands using Office 365 and Azure within months. And we did it! A challenging project and a perfect showcase to demonstrate the do’s and don’ts when migrating to the several components of Office 365, i.e. Exchange Online, SharePoint Online and OneDrive for Business at the speed of light.
In this session we will demonstrate how we:
- Analyzed the on-premises installations
- Cleaned up the obsolete users and data
- Prepared the Office 365-tenant to host 3000 users and over 30 domains
- Set-up the AD FS-infrastructure in Windows Azure
- Migrated 3000 users, their mail domains and aliases
- Migrated the file shares to OneDrive for Business and SharePoint Online
- Migrated the applications to SharePoint Online and Azure
- The PowerShell workarounds we used to fix several incompatibility issues between Office 365 and the on-premises installations…and what not to do.
Henry Peyret Presentation - Data Governance 2.0.
Based on the analysis of Digital Transformation and Values Transformation, Forrester gives its insight and orientations in terms of Data Governance 2.0 and Data Citizenship.
data management, information management, data, big data, personal organization, organization, file management, scientific research, research, project management, data security, file naming conventions, data management plan,
This describes a conceptual model approach to designing an enterprise data fabric. This is the set of hardware and software infrastructure, tools and facilities to implement, administer, manage and operate data operations across the entire span of the data within the enterprise across all data activities including data acquisition, transformation, storage, distribution, integration, replication, availability, security, protection, disaster recovery, presentation, analytics, preservation, retention, backup, retrieval, archival, recall, deletion, monitoring, capacity planning across all data storage platforms enabling use by applications to meet the data needs of the enterprise.
The conceptual data fabric model represents a rich picture of the enterprise’s data context. It embodies an idealised and target data view.
Designing a data fabric enables the enterprise respond to and take advantage of key related data trends:
• Internal and External Digital Expectations
• Cloud Offerings and Services
• Data Regulations
• Analytics Capabilities
It enables the IT function demonstrate positive data leadership. It shows the IT function is able and willing to respond to business data needs. It allows the enterprise to meet data challenges
• More and more data of many different types
• Increasingly distributed platform landscape
• Compliance and regulation
• Newer data technologies
• Shadow IT where the IT function cannot deliver IT change and new data facilities quickly
It is concerned with the design an open and flexible data fabric that improves the responsiveness of the IT function and reduces shadow IT.
Why apache Flink is the 4G of Big Data Analytics FrameworksSlim Baltagi
Apache Flink is a community-driven open source and memory-centric Big Data analytics framework. It provides the only hybrid (Real-Time Streaming + Batch) open source distributed data processing engine supporting many use cases.
Flink uses a mixture of Scala and Java internally, has very good Scala APIs and some of its libraries are basically pure Scala (FlinkML and Table).
At its core, it is a streaming dataflow execution engine and it also provides several APIs for batch processing (DataSet API), real-time streaming (DataStream API) and relational queries (Table API) and also domain-specific libraries for machine learning (FlinkML) and graph processing (Gelly).
In this talk, you will learn in more details about:
What is Apache Flink, how it fits into the Big Data ecosystem and why it is the 4G (4th Generation) of Big Data Analytics frameworks?
How Apache Flink integrates with Apache Hadoop and other open source tools for data input and output as well as deployment?
Why Apache Flink is an alternative to Apache Hadoop MapReduce, Apache Storm and Apache Spark? What are the benchmarking results between Apache Flink and those other Big Data analytics frameworks?
Overview of Data Loss Prevention Policies in Office 365Dock 365
Presentation about identifying, monitoring, and automatically protect sensitive information across Office 365.
With a DLP Policy, you can:
- Identify sensitive information across many locations, such as SharePoint Online and OneDrive for Business.
- Prevent the accidental sharing of sensitive information.
- Monitor and protect sensitive information in the desktop versions of Excel 2016, PowerPoint 2016, and Word 2016.
- Help users learn how to stay compliant without interrupting their workflow.
- View DLP reports showing content that matches your organization's DLP policies.
Visit www.mydock365.com to learn more about SharePoint with Dock.
Migrating 3000 users and 1100 applications from Lotus Notes to Office 365Arno Flapper
A large international customer decided to move into Microsoft Office 365. Their challenge is to migrate 3000+ users, 1100 applications and 100 Tb of file share data located in 50 locations and datacenters across Europe. With this migration all datacenters must be replaced by a central datacenter in The Netherlands using Office 365 and Azure within months. And we did it! A challenging project and a perfect showcase to demonstrate the do’s and don’ts when migrating to the several components of Office 365, i.e. Exchange Online, SharePoint Online and OneDrive for Business at the speed of light.
In this session we will demonstrate how we:
- Analyzed the on-premises installations
- Cleaned up the obsolete users and data
- Prepared the Office 365-tenant to host 3000 users and over 30 domains
- Set-up the AD FS-infrastructure in Windows Azure
- Migrated 3000 users, their mail domains and aliases
- Migrated the file shares to OneDrive for Business and SharePoint Online
- Migrated the applications to SharePoint Online and Azure
- The PowerShell workarounds we used to fix several incompatibility issues between Office 365 and the on-premises installations…and what not to do.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Which Change Data Capture Strategy is Right for You?Precisely
Change Data Capture or CDC is the practice of moving the changes made in an important transactional system to other systems, so that data is kept current and consistent across the enterprise. CDC keeps reporting and analytic systems working on the latest, most accurate data.
Many different CDC strategies exist. Each strategy has advantages and disadvantages. Some put an undue burden on the source database. They can cause queries or applications to become slow or even fail. Some bog down network bandwidth, or have big delays between change and replication.
Each business process has different requirements, as well. For some business needs, a replication delay of more than a second is too long. For others, a delay of less than 24 hours is excellent.
Which CDC strategy will match your business needs? How do you choose?
View this webcast on-demand to learn:
• Advantages and disadvantages of different CDC methods
• The replication latency your project requires
• How to keep data current in Big Data technologies like Hadoop
Unlocking Greater Insights with Integrated Data Quality for CollibraPrecisely
Data is arguably your company’s greatest asset, and a thoughtful data governance strategy, along with robust tools like Collibra Data Governance Center (DGC), is essential to getting the most value from that data. However, even the best data governance programs will falter without data quality.
Data governance systems provide a framework for the policies, processes, rules, roles and responsibilities that help you manage your enterprise data. But they don’t give you insight into the characteristics and quality of that data – such as errors, outliers and issues – nor how the data changes over time.
During this webinar, we discuss how seamlessly integrating Trillium DQ with Collibra DGC creates a complete data governance solution that delivers rapid insights into the health of your data, ensuring trust and compliance with organizational policies and plans. We demonstrate how data is automatically exchanged between the tools so users can:
• Quickly establish the rules needed to support policies
• Evaluate their data against those rules on an ongoing basis
• Identify problems or improvements with their data quality to take action
IRMS UG Principles of Retention in Microsoft 365Joanne Klein
There's only 4 of them, but they're important. Really important.
This session will talk about why it's essential for Records Managers and Information Management teams to have an authoritative understanding of the principles. They inform the retention and deletion processes and actions which are the technical underpinning of your Information Governance/Records Management (IG/RM) implementation. This is what ultimately determines the retention and deletion actions taken on each piece of content.
Understanding the principles of retention is an important prerequisite to a holistic IG/RM design and implementation. It can help ensure you are deleting content when you no longer need/require it and retaining content for as long as you are required to meet your corporate governance requirements and your legal, business, and regulatory obligations.
More and more frequently I need to explain to people the concept of Social Media Intelligence (SOCMINT).
This is a simple slidedeck to introduce people the new concept and its value in digital marketing and strategy and what is the power behind it.
Download at http://DavidHubbard.net/powerpoint - This Introduction to Business Intelligence gives an overview of how Business Intelligence fits into business strategy in general. It does not go into the specific technologies of Business Intelligence. It is meant to be used to explain Business Intelligence to those not already familiar with Business Intelligence.
The data services marketplace is enabled by a data abstraction layer that supports rapid development of operational applications and single data view portals. In this presentation yo will learn services-based reference architecture, modality, and latency of data access.
- Reference architecture for enterprise data services marketplace
- Modality and latency of data access
- Customer use cases and demo
This presentation is part of the Denodo Educational Seminar , and you can watch the video here goo.gl/vycYmZ.
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Spline: Apache Spark Lineage not Only for the Banking Industry with Marek Nov...Spark Summit
Data lineage tracking is one of the significant problems that financial institutions face when using modern big data tools. This presentation describes Spline – a data lineage tracking and visualization tool for Apache Spark. Spline captures and stores lineage information from internal Spark execution plans and visualizes it in a user-friendly manner.
DataOps - Big Data and AI World London - March 2020 - Harvinder AtwalHarvinder Atwal
Title
DataOps, the secret weapon for delivering AI, data science, and business intelligence value at speed.
Synopsis
● According to recent research, just 7.3% of organisations say the state of their data and analytics is excellent, and only 22% of companies are currently seeing a significant return from data science expenditure.
● Poor returns on data & analytics investment are often the result of applying 20th-century thinking to 21st-century challenges and opportunities.
● Modern data science and analytics require secure, efficient processes to turn raw data from multiple sources and in numerous formats into useful inputs to a data product.
● Developing, orchestrating and iterating modern data pipelines is an extremely complex process requiring multiple technologies and skills.
● Other domains have to successfully overcome the challenge of delivering high-quality products at speed in complex environments. DataOps applies proven agile principles, lean thinking and DevOps practices to the development of data products.
● A DataOps approach aligns data producers, analytical data consumers, processes and technology with the rest of the organisation and its goals.
Data Architecture Best Practices for Today’s Rapidly Changing Data LandscapeDATAVERSITY
With the rise of the data-driven organization, the pace of innovation in data-centric technologies has been tremendous. New tools and techniques are emerging at an exponential rate, and it is difficult to keep track of the array of technological choices available to today’s data management professional.
At the same time, core fundamentals such as data quality and metadata management remain critical in order for organizations to obtain true business value from their data. This webinar will help demystify the options available: from data lake to data warehouse, to graph database, to NoSQL, and more, and how to integrate these new technologies with core architectural fundamentals that will help your organization benefit from the quick wins that are possible from these exciting technologies, while at the same time build a longer-term sustainable architecture that will support the inevitable change that will continue in the industry.
An AI Maturity Roadmap for Becoming a Data-Driven OrganizationDavid Solomon
The initial version of a maturity roadmap to help guide businesses when adopting AI technology into their workflow. IBM Watson Studio is referenced as an example of technology that can help in accelerating the adoption process.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Which Change Data Capture Strategy is Right for You?Precisely
Change Data Capture or CDC is the practice of moving the changes made in an important transactional system to other systems, so that data is kept current and consistent across the enterprise. CDC keeps reporting and analytic systems working on the latest, most accurate data.
Many different CDC strategies exist. Each strategy has advantages and disadvantages. Some put an undue burden on the source database. They can cause queries or applications to become slow or even fail. Some bog down network bandwidth, or have big delays between change and replication.
Each business process has different requirements, as well. For some business needs, a replication delay of more than a second is too long. For others, a delay of less than 24 hours is excellent.
Which CDC strategy will match your business needs? How do you choose?
View this webcast on-demand to learn:
• Advantages and disadvantages of different CDC methods
• The replication latency your project requires
• How to keep data current in Big Data technologies like Hadoop
Unlocking Greater Insights with Integrated Data Quality for CollibraPrecisely
Data is arguably your company’s greatest asset, and a thoughtful data governance strategy, along with robust tools like Collibra Data Governance Center (DGC), is essential to getting the most value from that data. However, even the best data governance programs will falter without data quality.
Data governance systems provide a framework for the policies, processes, rules, roles and responsibilities that help you manage your enterprise data. But they don’t give you insight into the characteristics and quality of that data – such as errors, outliers and issues – nor how the data changes over time.
During this webinar, we discuss how seamlessly integrating Trillium DQ with Collibra DGC creates a complete data governance solution that delivers rapid insights into the health of your data, ensuring trust and compliance with organizational policies and plans. We demonstrate how data is automatically exchanged between the tools so users can:
• Quickly establish the rules needed to support policies
• Evaluate their data against those rules on an ongoing basis
• Identify problems or improvements with their data quality to take action
IRMS UG Principles of Retention in Microsoft 365Joanne Klein
There's only 4 of them, but they're important. Really important.
This session will talk about why it's essential for Records Managers and Information Management teams to have an authoritative understanding of the principles. They inform the retention and deletion processes and actions which are the technical underpinning of your Information Governance/Records Management (IG/RM) implementation. This is what ultimately determines the retention and deletion actions taken on each piece of content.
Understanding the principles of retention is an important prerequisite to a holistic IG/RM design and implementation. It can help ensure you are deleting content when you no longer need/require it and retaining content for as long as you are required to meet your corporate governance requirements and your legal, business, and regulatory obligations.
More and more frequently I need to explain to people the concept of Social Media Intelligence (SOCMINT).
This is a simple slidedeck to introduce people the new concept and its value in digital marketing and strategy and what is the power behind it.
Download at http://DavidHubbard.net/powerpoint - This Introduction to Business Intelligence gives an overview of how Business Intelligence fits into business strategy in general. It does not go into the specific technologies of Business Intelligence. It is meant to be used to explain Business Intelligence to those not already familiar with Business Intelligence.
The data services marketplace is enabled by a data abstraction layer that supports rapid development of operational applications and single data view portals. In this presentation yo will learn services-based reference architecture, modality, and latency of data access.
- Reference architecture for enterprise data services marketplace
- Modality and latency of data access
- Customer use cases and demo
This presentation is part of the Denodo Educational Seminar , and you can watch the video here goo.gl/vycYmZ.
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Spline: Apache Spark Lineage not Only for the Banking Industry with Marek Nov...Spark Summit
Data lineage tracking is one of the significant problems that financial institutions face when using modern big data tools. This presentation describes Spline – a data lineage tracking and visualization tool for Apache Spark. Spline captures and stores lineage information from internal Spark execution plans and visualizes it in a user-friendly manner.
DataOps - Big Data and AI World London - March 2020 - Harvinder AtwalHarvinder Atwal
Title
DataOps, the secret weapon for delivering AI, data science, and business intelligence value at speed.
Synopsis
● According to recent research, just 7.3% of organisations say the state of their data and analytics is excellent, and only 22% of companies are currently seeing a significant return from data science expenditure.
● Poor returns on data & analytics investment are often the result of applying 20th-century thinking to 21st-century challenges and opportunities.
● Modern data science and analytics require secure, efficient processes to turn raw data from multiple sources and in numerous formats into useful inputs to a data product.
● Developing, orchestrating and iterating modern data pipelines is an extremely complex process requiring multiple technologies and skills.
● Other domains have to successfully overcome the challenge of delivering high-quality products at speed in complex environments. DataOps applies proven agile principles, lean thinking and DevOps practices to the development of data products.
● A DataOps approach aligns data producers, analytical data consumers, processes and technology with the rest of the organisation and its goals.
Data Architecture Best Practices for Today’s Rapidly Changing Data LandscapeDATAVERSITY
With the rise of the data-driven organization, the pace of innovation in data-centric technologies has been tremendous. New tools and techniques are emerging at an exponential rate, and it is difficult to keep track of the array of technological choices available to today’s data management professional.
At the same time, core fundamentals such as data quality and metadata management remain critical in order for organizations to obtain true business value from their data. This webinar will help demystify the options available: from data lake to data warehouse, to graph database, to NoSQL, and more, and how to integrate these new technologies with core architectural fundamentals that will help your organization benefit from the quick wins that are possible from these exciting technologies, while at the same time build a longer-term sustainable architecture that will support the inevitable change that will continue in the industry.
An AI Maturity Roadmap for Becoming a Data-Driven OrganizationDavid Solomon
The initial version of a maturity roadmap to help guide businesses when adopting AI technology into their workflow. IBM Watson Studio is referenced as an example of technology that can help in accelerating the adoption process.
Falcon.io | 2021 Trends Virtual Summit - Data PrivacyFalcon.io
Two years after Cambridge Analytica, personal data is still at the forefront of the conversation. Consumers all vehemently care about what companies do with their data. Privacy is on everyone’s mind (not just the older, wealthier bracket). The pandemic has only accelerated these concerns, which marked 2020 as a significant year for data management. As businesses face new technological challenges amid COVID-19, the topic of privacy has drawn an increasingly bright spotlight. This sensitivity regarding data privacy has only been growing — and companies are responding accordingly.
Art Hathaway - Artificial Intelligence - Real Threat Preventioncentralohioissa
Throughout history we've seen opposing forces skillfully pit strengths against weaknesses until, ultimately, one side succumbs. Holding a position takes considerably more effort than does a single, offensive surge, and attackers are counting on it. The very nature of the cybersecurity attacks we face today are in direct response to the shortcomings of the available tools, knowledge and approaches. The only problem is that we must evolve our defenses as fast as (or faster) than their offenses, and the odds are greatly in their favor. Imagine a football game – with no time limits – determined by your opponent’s first undefended scoring play. Game over. Hmmm…I wonder how that one ends?
Facing next-generation challenges requires a next-generation approach – preferably one that requires no change to your current production environment, never tires, continually evolves, doesn't rely on humans and is 99%+ accurate regardless of Internet connectivity. We'll discuss a solution that shifts the balance in your favor by leveraging artificial intelligence to predict and prevent against malware-born threats so you don't have to.
10 Steps for Taking Control of Your Organization's Digital Debris Perficient, Inc.
Do you have too much old information, but not enough guidance to begin the task of cleaning out your data stores? Join Perficient to learn 10 tips for creating a strategic roadmap to take control of your information and uncover the technology that can support your efforts, including how to:
Stop keeping everything forever
Create an information governance and disposal policy before implementing technology
Automate information management to improve employee productivity
Prepare a discovery response plan
How to Create 80% of a Big Data Pilot ProjectGreg Makowski
When evaluating Open Source Software, or other software of a certain size or complexity, organizations frequently want to conduct a Pilot project, or Proof of Concept (POC). This talk describes a process to reduce the length of the Pilot, by leveraging configurations from performance testing to POC starting configurations.
Algorithm Marketplace and the new "Algorithm Economy"Diego Oppenheimer
Talk by Diego Oppenheimer CEO of Algorithmia.com at Data Day Texas 2016.
Peter Sondergaard VP of Research for Gartner recently said the next digital gold rush is "How we do something with data not just what you do with it". During this talk we will cover a brief history of the different algorithmic advances in computer vision, natural language processing, machine learning and general AI and how they are being applied to Big Data today. From there we will talk about how algorithms are playing a crucial part in the next Big Data revolution, new opportunities that are opening up for startups and large companies alike as well as a first look into the role Algorithm Marketplaces will play in this space.
ADV Slides: Increasing Artificial Intelligence Success with Master Data Manag...DATAVERSITY
Companies all over the world are going through a digital transformation now, which in many cases, is all about maturing the data environment and the use of data. Master data is key to this effort. All transformative projects require master data and usually many subject areas. Current efforts to deliver master data to the enterprise are cumbersome, inefficient, and met with limited acceptance.
We’ll look at enterprise use cases of artificial intelligence and show the master data that is needed. We’ll see what some MDM vendors are doing with AI and how the future of MDM will be shaped by looking at some specific MDM actions influenced by AI.
apidays LIVE Hong Kong 2021 - Federated Learning for Banking by Isaac Wong, W...apidays
apidays LIVE Hong Kong 2021 - API Ecosystem & Data Interchange
August 25 & 26, 2021
Federated Learning for Banking
Isaac Wong, AI Solution Architect at WeBank
2013 Data Protection Maturity Trends: How Do You Compare?Lumension
In 2012 we found out that the BYOD environment and consumerization of the workplace had turned traditional notions of corporate IT upside down. The 2013 Data Protection Maturity Report will highlight how organizations have managed this trend over the last year and what steps are being taken in 2013 to further enhance data security. Find out how IT teams are developing a holistic model that encompasses policy, education, technology and enforcement.
Within this slide deck, we look at each of data protection trends, helping you define your organization’s best practice guide to address the top concerns. We will also be showing you how you can gauge the maturity of your security systems, allowing you to plug any holes before your valuable data starts to leak through them.
Similar to How Walmart is Combating Fraud and Saving Consumers Millions (20)
An introduction to Elasticsearch's advanced relevance ranking toolboxElasticsearch
The hallmark of a great search experience is always delivering the most relevant results, quickly, to every user. The difficulty lies behind the scenes in making that happen elegantly and at a scale. From App Search’s intuitive drag and drop interface to the advanced relevance capabilities built into the core of Elasticsearch — Elastic offers a range of tools for developers to tune relevance ranking and create incredible search experiences. In this session, we’ll explore some of Elasticsearch’s advanced relevance ranking features, such as dense vector fields, BM25F, ranking evaluation, and more. Plus we’ll give you some ideas for how these features are being used by other Elastic users to create world-class, category defining search experiences.
Eze Castle Integration is a managed service provider (MSP), cloud service provider (CSP), and internet service provider (ISP) that delivers services to more than 1,000 clients around the world. Different departments within Eze Castle have devised their own log aggregation solutions in order to provide visibility, meet regulatory compliance requirements, conduct cybersecurity investigations, and help engineers with troubleshooting infrastructure issues. In 2019, they partnered with Elastic to consolidate the data generated from different systems into a single pane of glass. And thanks to the ease of deployment on Elastic Cloud, professional consultation services from Elastic engineers, and on-demand training courses available on Elastic Learning, Eze Castle was able to go from proof-of-concept to a fully functioning ""Eze Managed SIEM"" product within a month!
Learn about Eze Castle's journey with Elastic and how they grew Eze Managed SIEM from zero to 100 customers In less than 14 months.
Cómo crear excelentes experiencias de búsqueda en sitios webElasticsearch
Descubre lo fácil que es crear búsquedas relevantes y enriquecidas en sitios web de cara al público para impulsar las conversiones, incrementar el consumo de contenido y ayudar a los visitantes a encontrar lo que necesitan. Realiza un recorrido por las herramientas de Elastic a las que puedes sacar partido para transformar con facilidad tu sitio web, lo que incluye nuestro nuevo y potente rastreador web.
Te damos la bienvenida a una nueva forma de realizar búsquedas Elasticsearch
Al igual que la mayoría de las organizaciones modernas, tus equipos probablemente usan más de 10 aplicaciones basadas en la nube a diario, pero dedican demasiado tiempo a buscar la información que necesitan en todas estas. Gracias a las características integradas de Elastic Workplace Search, podrás comprobar lo sencillo que resulta poner el contenido relevante al alcance de tus equipos gracias a la búsqueda unificada para todas las aplicaciones que usan para llevar a cabo su trabajo.
Tirez pleinement parti d'Elastic grâce à Elastic CloudElasticsearch
Découvrez pourquoi Elastic Cloud est la solution idéale pour exploiter toutes les offres d'Elastic. Bénéficiez d'une flexibilité d'achat et de déploiement au sein de Google Cloud, de Microsoft Azure, d'Amazon Web Services ou des trois à la fois. Apprenez quels avantages vous apporte une offre de service géré et déterminez la solution qui vous permet de la gérer par vous-même grâce à des outils intégrés d'automatisation et d'orchestration. Et ce n'est pas tout ! Familiarisez-vous avec les fonctionnalités qui peuvent vous aider à scaler vos opérations au fur et à mesure de l'évolution de votre déploiement, à stocker vos données d'une manière rentable et à optimiser vos recherches. Ainsi, vous n'aurez plus à abandonner de données et obtiendrez les informations exploitables dont vous avez besoin pour assurer le fonctionnement de votre entreprise.
Comment transformer vos données en informations exploitablesElasticsearch
Découvrez des fonctionnalités stratégiques de la Suite Elastic, notamment Elasticsearch, un moteur de données incomparable, et Kibana, véritable fenêtre ouverte sur la Suite Elastic.
Dans cette session, vous apprendrez à :
injecter des données dans la Suite Elastic ;
stocker des données ;
analyser des données ;
exploiter des données.
Plongez au cœur de la recherche dans tous ses états.Elasticsearch
À l'instar de la plupart des entreprises modernes, vos équipes utilisent probablement plus de 10 applications hébergées dans le cloud chaque jour, mais passent aussi bien trop de temps à chercher les informations dont elles ont besoin dans ces outils. Grâce aux fonctionnalités prêtes à l'emploi d'Elastic Workplace Search, découvrez combien il est facile de mettre le contenu pertinent à portée de la main de vos équipes grâce à une recherche unifiée sur l'ensemble des applications qu'elles utilisent pour faire leur travail.
Modernising One Legal Se@rch with Elastic Enterprise Search [Customer Story]Elasticsearch
Knowledge management needs in the legal sector, why Linklaters decided to move away from its legacy KM search engine, Kin+Carta's management of the migration process, and how the switch revitalised a well-established system and opened up new possibilities for its future development.
An introduction to Elasticsearch's advanced relevance ranking toolboxElasticsearch
The hallmark of a great search experience is always delivering the most relevant results, quickly, to every user. The difficulty lies behind the scenes in making that happen elegantly and at a scale. From App Search’s intuitive drag and drop interface to the advanced relevance capabilities built into the core of Elasticsearch — Elastic offers a range of tools for developers to tune relevance ranking and create incredible search experiences. In this session, we’ll explore some of Elasticsearch’s advanced relevance ranking features, such as dense vector fields, BM25F, ranking evaluation, and more. Plus we’ll give you some ideas for how these features are being used by other Elastic users to create world-class, category defining search experiences.
Like most modern organizations, your teams are likely using upwards of 10 cloud-based applications on a daily basis, but spending far too many hours a day searching for the information they need across all of them. With the out-of-the-box capabilities of Elastic Workplace Search, see how easy it is to put relevant content right at your teams’ fingertips with unified search across all the apps they rely on to get work done.
Building great website search experiencesElasticsearch
Discover how easy it is to create rich, relevant search on public facing websites that drives conversion, increases content consumption, and helps visitors find what they need. Get a tour of the Elastic tools you can leverage to easily transform your website, including our powerful new web crawler.
Keynote: Harnessing the power of Elasticsearch for simplified searchElasticsearch
Get an overview of the innovation Elastic is bringing to the Enterprise Search landscape, and learn how you can harness these capabilities across your technology landscape to make the power of search work for you.
Cómo transformar los datos en análisis con los que tomar decisionesElasticsearch
Descubre las áreas de características estratégicas de Elastic Stack: Elasticsearch, un motor de datos inigualable y Kibana, la ventana que da acceso a Elastic Stack.
En la sesión hablaremos sobre:
Cómo incorporar datos a Elastic Stack
Almacenamiento de datos
Análisis de los datos
Actuar en función de los datos
Explore relève les défis Big Data avec Elastic Cloud Elasticsearch
Spécialisée dans le développement et la gestion de solutions de veille documentaire et commerciale, Explore offre à ses clients une lecture précise et organisée de l’actualités des marchés et projets sur leurs territoires d'intervention. Afin de rendre leur offre plus agile et performante, Explore a choisi l’offre Elastic Cloud hébergée sur Microsoft Azure. Découvrez comment les équipes de production et de développement sont désormais en mesure de mieux exploiter les données pour les clients d’Explore et gagnent du temps sur la gestion de leur infrastructure.
Comment transformer vos données en informations exploitablesElasticsearch
Découvrez des fonctionnalités stratégiques de la Suite Elastic, notamment Elasticsearch, un moteur de données incomparable, et Kibana, véritable fenêtre ouverte sur la Suite Elastic.
Dans cette session, vous apprendrez à :
injecter des données dans la Suite Elastic ;
stocker des données ;
analyser des données ;
exploiter des données.
Transforming data into actionable insightsElasticsearch
Learn about the strategic feature areas of the Elastic Stack—Elasticsearch, a data engine like no other, and Kibana, the window into the Elastic Stack.
The session will cover:
Bringing data into the Elastic Stack
Storing data
Analyzing data
Acting on data
"Elastic enables the world’s leading organization to exceed their business objectives and power their mission-critical systems by eliminating data silos, connecting the dots, and transforming data of all types into actionable insights.
Come learn how the power of search can help you quickly surface relevant insights at scale. Whether you are an executive looking to reduce operational costs, a department head striving to do more with fewer tools, or engineer monitoring and protecting your IT environment, this session is for you. "
Empowering agencies using Elastic as a Service inside GovernmentElasticsearch
It has now been four years since the beta release of Elastic Cloud Enterprise which kicked off a wave of the Elastic public sector community running Elastic as a service within Government rather than utilizing purely hosted solutions. Fast forward to 2021 and we have multiple options for multiple mission needs. Learn top tips from Elastic architects and their experience enabling their teams with the automation and provisioning of Elastic tech to change the game in how government delivers solutions.
The opportunities and challenges of data for public goodElasticsearch
Data is an increasingly valuable resource for delivering economic and social benefit. Heather will discuss the challenges and opportunities, and how communities at all levels of the public sector can play a part in leading the change.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
2. 2
“We want to foster a culture that supports change and
helps us go along the journey we’re all on, which will
lead to even more change in the future”
- Doug McMillon
“
3. 3
Data Science – Team Size 5
Industrial Engineer
Economist
Software/ Web Developer
Info Systems Engineer
Statistician
Feeding Machines... Feeding People...
Making Data Accessible...
Data Analytics- Team Size 30
Mathematicians
Crime Analysts
Accountants
eCommerce Analysts
Global Media Analysts
Fred Helm
- Senior Director IAI
- 20 year in state and federal law
enforcement
- Developed IROC to action
Walmart risk in real time
Tim Fenton
- 14 patents disclosed with
Walmart
- Software engineer by background
- Data Scientist for past 3 years
- Worked previously at Motorola
and in US Intel agency
4. Varying backgrounds lead to incredible ideas and world-class
solutions
4
Know Your
Customer
Implementati
on
Content
Strategy
Concept
& Idea
Data
Analysis
Consultation
5. Walmart Global Investigations
5
Corporate
Ethics
Crimes Against
the Business
Organized
Retail Crime
Consumer
Frauds
Technology
Data
Exfiltration
eCommerce
Fraud
Health &
Wellness
Theft and
Fraud related
to meds
Insurance
Fraud
6. 6
• Proofpoint Logs/ metadata
• Email
• Chat Messages
• Contracts/ Agreements/ Statements
of Work
Communications ESI
Real-Time
• Think Teradata, Retail Link,
SAP
• Human Resources Data
• Travel & Expense
Cold Storage Databases
• System Logins
• Internet Searches
• Badge Entry/ Exit
Logs
Systems ESI
• Information provided through the
allegation
• Data available through front-end
interface
Allegation
9. GI Data Sciences: Skunk Works
Engines – T2
Buckeye
Landing Gear
– F5 Tiger
Tooling – C5
Galaxy
Flight
Control
System – F16
Falcon
“Have Blue”
Case Study In Re-Use:
• Lockheed “Have Blue” prototype
was about testing stealth
• Make use of off the shelf system
• In a similar vein GI DS team re-
uses Walmart IT components
11. Consumer Gift Card Fraud
Walmart activates billions in Gift Cards
yearly Financial processing through a 3rd party
Social engineering and gift card
activation
Customer Trust Reputation Compliance
12. ES Architecture for Gift Cards
Cluster
Node 1 Node 2 Node 3
Node 4 Node 5 Node 6
Node 7 Node 8
Node 10 Node 11
Node 9
Node 12
Kafka
Topic:
Point of
Sale
Kafka
Topic:
Gift Card
Balance
Checks
Spring Boot
App
Filtered Transactions:
Peak 5,000/sec
Logstash
Filtered Transactions:
Peak 50/sec
Runs in
Internal Cloud:
VM’s 110GB
RAM, 1.1TB
Disk
Adjudicate
Gift Cards
Every X
Seconds
Logstash for
archive/ML
training
15. Gift Card Graph
- Formed from 1
suspicious customer id
- Linked another
suspicious customer via
non-US IP
- Allowed us to action
compromised cards
16. Results
• Millions in customer dollars protected
• Making fraud harder to commit
• Able to automate a 24/7 protection system w/out humans in
the loop